HELP

AI-900 Mock Exam Marathon: Timed Sims & Weak Spot Repair

AI Certification Exam Prep — Beginner

AI-900 Mock Exam Marathon: Timed Sims & Weak Spot Repair

AI-900 Mock Exam Marathon: Timed Sims & Weak Spot Repair

Timed AI-900 sims plus targeted drills to fix weak areas fast.

Beginner ai-900 · microsoft · azure-ai-fundamentals · ai-certification

Course purpose: AI-900 practice that feels like the real exam

This course is a focused, simulation-first blueprint for learners preparing for Microsoft’s AI-900: Azure AI Fundamentals exam. Instead of long lectures, you’ll use timed mock exams and targeted weak-spot repair to build confidence with the official objectives and the way Microsoft asks questions. The result: faster recall, better service selection, and fewer avoidable mistakes under time pressure.

You’ll work through a 6-chapter “book” structure. Chapter 1 sets you up with the exam mechanics (registration, delivery options, scoring expectations, and pacing strategy). Chapters 2–5 align directly to the official AI-900 exam domains and include exam-style drills designed to mirror real objective wording. Chapter 6 delivers a full mock exam experience and a structured final review loop.

If you’re new to certifications, you’ll also learn how to read Microsoft-style prompts, eliminate distractors, and avoid common traps (like confusing similar service families or picking an overly complex solution).

Aligned to official AI-900 exam domains

This blueprint maps to the published AI-900 domains and keeps practice anchored to objective names:

  • Describe AI workloads: identify when AI is appropriate, recognize common AI workload types, and apply responsible AI basics.
  • Fundamental principles of ML on Azure: understand supervised vs. unsupervised learning, core model lifecycle concepts, and evaluation at a fundamentals level.
  • Computer vision workloads on Azure: choose capabilities for image analysis and OCR-style scenarios.
  • NLP workloads on Azure: map language scenarios to Azure AI Language and conversational patterns.
  • Generative AI workloads on Azure: explain generative AI concepts and apply Azure OpenAI Service usage patterns responsibly.

How the “Mock Exam Marathon” method works

Each domain chapter includes deep explanation at the level expected for AI-900 and then pivots quickly into exam-style practice. After each drill set, you’ll do a short review loop: (1) identify which objective the question tested, (2) record why your chosen answer was tempting, and (3) write a one-sentence rule you can reuse on the next similar prompt.

Chapter 6 brings everything together with two timed simulations and a structured weak-spot analysis process. You’ll end with a last-pass checklist that covers pacing, reviewing flagged questions, and keeping your choices aligned to “fundamentals” scope (a common pitfall is selecting services or techniques beyond what the exam expects).

Who this course is for

This course is designed for beginners with basic IT literacy and no prior certification experience. You don’t need to be a developer or data scientist. You’ll learn to recognize scenarios and pick the most appropriate Azure AI approach—exactly what AI-900 tests.

Get started on Edu AI

Create your learning plan and start practicing today. Register free to track progress and revisit weak domains, or browse all courses to compare certification paths.

Outcome

By the end, you’ll have completed domain-aligned drills, two full timed simulations, and a repeatable weak-spot repair routine—so you can walk into the Microsoft AI-900 exam with realistic timing practice and clear decision rules for every domain.

What You Will Learn

  • Describe AI workloads: identify common AI scenarios, approaches, and responsible AI considerations
  • Fundamental principles of ML on Azure: choose ML types, core concepts, and Azure Machine Learning options
  • Computer vision workloads on Azure: select Azure AI Vision capabilities for image analysis and OCR use cases
  • NLP workloads on Azure: map language tasks to Azure AI Language services and conversational AI concepts
  • Generative AI workloads on Azure: explain gen AI concepts and apply Azure OpenAI Service patterns responsibly
  • Exam readiness: manage timing, interpret tricky wording, and use weak-spot analytics to raise your score

Requirements

  • Basic IT literacy (web apps, files, browsers, and cloud concepts at a high level)
  • No prior certification experience required
  • A computer with reliable internet access
  • Optional: an Azure account for exploring services (not required for passing this course)

Chapter 1: AI-900 Exam Orientation and Study Game Plan

  • Understand AI-900 format, timing, and question styles
  • Register for the exam and set a realistic schedule
  • Build a study plan aligned to official domains
  • Set up your practice environment and baseline assessment

Chapter 2: Describe AI Workloads (Domain 1) — Scenarios to Services

  • Recognize AI workloads and when AI is appropriate
  • Match workload types to Azure AI services at a high level
  • Apply responsible AI principles in exam scenarios
  • Timed domain drill set: mixed single/multiple-answer questions

Chapter 3: Fundamental Principles of ML on Azure (Domain 2)

  • Master ML fundamentals: features, labels, training, and inference
  • Choose the right ML type for a given scenario
  • Understand evaluation metrics and overfitting at exam depth
  • Timed simulation: ML-focused caselets and MCQs

Chapter 4: Computer Vision Workloads on Azure (Domain 3)

  • Identify vision tasks: image analysis, OCR, and detection
  • Map tasks to Azure AI Vision capabilities
  • Understand performance considerations and responsible use cases
  • Timed simulation: vision-heavy items and service selection

Chapter 5: NLP + Generative AI Workloads on Azure (Domains 4–5)

  • Select NLP solutions for classification, extraction, Q&A, and conversation
  • Understand conversational AI building blocks and orchestration
  • Explain generative AI concepts, prompting, and grounding on Azure
  • Timed simulation: mixed NLP + generative AI objective questions

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist

Jordan McAllister

Microsoft Certified Trainer (MCT)

Jordan McAllister is a Microsoft Certified Trainer who builds beginner-friendly certification prep for Azure fundamentals tracks. He specializes in translating Microsoft exam objectives into focused practice and repeatable test-taking strategies for first-time candidates.

Chapter 1: AI-900 Exam Orientation and Study Game Plan

AI-900 is a fundamentals exam, but “fundamentals” in Microsoft exams still means you must make consistent, test-aligned decisions: identify the workload, choose the right Azure service category, and recognize responsible AI considerations. This course is built as a Mock Exam Marathon, so your job in Chapter 1 is to set up the rules of the game: understand the format and policies, build a schedule that prevents cram-burnout, and establish a baseline so practice time targets your weak spots instead of repeating what you already know.

Across this book, you’ll map common scenarios to the official domains: describing AI workloads (including responsible AI), machine learning principles on Azure, computer vision workloads, NLP workloads, and generative AI workloads with Azure OpenAI patterns. The exam rewards clarity: if you can name the task (classification, OCR, entity extraction, chat completion) and match it to the correct service family, you will score well. The rest is execution: timing, reading precision, and a disciplined review loop.

Exam Tip: Treat AI-900 like a vocabulary-and-decision exam. When you miss a question, don’t just note the right option—write the rule that would have made the choice obvious (e.g., “OCR → Azure AI Vision Read,” “custom model training → Azure Machine Learning,” “chatbot orchestration → Azure AI Bot Service + Language/Orchestrator,” “generative text → Azure OpenAI”).

This chapter walks you through exam orientation and a study game plan aligned to official objectives, then sets up your practice environment and baseline assessment so your timed sims are diagnostic, not just repetitive drilling.

Practice note for Understand AI-900 format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set a realistic schedule: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a study plan aligned to official domains: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Set up your practice environment and baseline assessment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand AI-900 format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set a realistic schedule: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a study plan aligned to official domains: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Set up your practice environment and baseline assessment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: What AI-900 Azure AI Fundamentals measures

AI-900 measures whether you can recognize common AI scenarios and select the appropriate Azure AI approach at a high level. It is not a coding exam, but it is also not purely definitional. You will repeatedly be asked to interpret a short business scenario, identify the AI workload type, and pick the best Azure service category or concept. The core domains you should map your study plan to are: (1) describing AI workloads and responsible AI, (2) fundamental ML principles on Azure (types of ML, training vs inference, features/labels, evaluation), (3) computer vision workloads (image analysis, object detection concepts, OCR), (4) NLP workloads (sentiment, key phrases, entity recognition, translation, conversation), and (5) generative AI workloads (LLMs, prompts, completions, embeddings, responsible use).

What the exam “tests for” is judgment: can you choose the most direct capability for the task? For example, if the requirement is to extract text from an image, your mental path should be “computer vision → OCR → Azure AI Vision.” If the requirement is to build and train a custom model with experiments, compute, datasets, and pipelines, think “Azure Machine Learning.” If the requirement is to analyze user reviews for sentiment or extract entities, think “Azure AI Language.” For generative scenarios, you should recognize when the task is content generation or summarization (Azure OpenAI), versus when it is classic NLP classification/extraction (Azure AI Language).

Exam Tip: Practice translating the scenario into a single verb-noun task statement before looking at options: “classify images,” “extract printed text,” “summarize emails,” “identify PII,” “predict house prices.” That one line often reveals the correct service family.

A frequent trap is confusing “AI workloads” with “Azure products.” The exam expects you to first name the workload (vision, language, ML, gen AI) and only then select the best-fitting service. If your approach starts with scanning for product names, you’ll be pulled into distractors that sound plausible but don’t match the requirement.

Section 1.2: Registration, delivery options, and exam policies

Your score improves when logistics stop being a variable. Register early, pick a date that supports spaced repetition, and choose a delivery option (test center vs online proctored) that matches your environment. Online proctoring is convenient but unforgiving: unstable internet, dual-monitor setups, background noise, and unapproved materials can end a session. Test center delivery reduces home-environment risk but adds travel and scheduling constraints. Either way, plan your study schedule backward from the exam date with milestones: baseline diagnostic, domain study blocks, timed sims, and final weak-spot repair.

Policies matter because they affect your performance psychology. You should know what identification is required, when you can reschedule, and how check-in works. If you choose online proctoring, do a device and network check well before exam day, and prepare a clean workspace. If you choose a test center, confirm location, arrival time, and what you can bring. The goal is to prevent “surprise stress” that steals time and focus.

Exam Tip: Book your exam for a time of day when you reliably do focused reading. AI-900 questions are short, but the trick is subtle wording; you want your best attention, not your last energy.

Build a realistic schedule: for many learners, 2–4 weeks with consistent practice is better than a single intense weekend. Allocate time across all domains even if you have prior experience—fundamentals exams often include breadth questions that punish gaps (for example, responsible AI principles or the difference between classic NLP and generative AI patterns).

Section 1.3: Scoring, cut score concept, and retake strategy

Microsoft exams typically report scores on a scaled range, and the “cut score” is the minimum passing threshold. You don’t need to chase perfection; you need dependable accuracy across domains. The practical implication is that your study plan should aim for balanced competence, not peak performance in one area and weakness in another. In fundamentals exams, a few missed questions in a small domain can be the difference between pass and fail if you ignored that topic entirely.

Because you won’t see “partial credit” in the way you study, train yourself to make clean decisions. When reviewing practice results, categorize misses into three buckets: (1) concept gap (you didn’t know the definition or service), (2) misread (you missed a keyword like “real-time,” “custom,” “translated,” “handwritten”), and (3) overthinking (you changed from the correct basic choice to an advanced-sounding distractor). Your retake strategy—ideally never needed—depends on which bucket dominates.

Exam Tip: Track your “avoidable errors” rate (misreads + overthinking). If that rate is high, more content study won’t fix it; you need timed reading drills and a stricter answer-selection process.

If you do need a retake, treat it as a targeted repair cycle: re-run a baseline diagnostic, focus on weak domains, and do timed simulations that mirror the real constraint. Avoid taking the exam again with the same approach; change the system (study loop, timing strategy, error log) rather than simply adding hours.

Section 1.4: How to use timed simulations and review loops

This course is a Mock Exam Marathon: your main training tool is timed simulations (“timed sims”) followed by disciplined review loops. The timed sim builds the two things content reading cannot: pacing and decision-making under pressure. Start with a baseline timed sim early—even if you feel unprepared—because it reveals which domains and question styles cost you time. Then use short, frequent sims (or timed sets) to practice reading precision and to stabilize your approach.

Your review loop is where score gains happen. After each sim, do not merely check correct answers. Instead, write an error log entry with: the workload type, the key requirement word(s), the correct service/concept, and the reason your choice was wrong. Over time, you’ll see patterns: confusing OCR with image analysis, mixing Azure AI Language with Azure OpenAI, or forgetting responsible AI principles. Convert those patterns into “one-line rules” you can recall under time pressure.

  • Run a timed sim to completion (avoid pausing; train realism).
  • Review wrong answers first, then “guessed correct” answers (these are fragile knowledge).
  • Extract 3–5 rules and add them to a short daily review list.
  • Re-run a focused timed set on the weakest objective area within 48 hours.

Exam Tip: When a question feels unfamiliar, anchor to the task type and the simplest matching service. Fundamentals exams reward direct mapping more than architectural creativity.

Use your practice environment to reinforce concepts: explore Azure AI service names and what they do at a high level, but don’t fall into the trap of trying to “learn everything in the portal.” Your goal is scenario-to-solution mapping, not implementation mastery.

Section 1.5: Common traps in Microsoft objective-style questions

Microsoft objective-style questions are engineered with plausible distractors. The most common trap is “almost right, but not for this requirement.” Train yourself to spot requirement keywords and eliminate options that violate them. Examples of requirement keywords include: “custom model,” “no code,” “real-time,” “extract text,” “detect faces,” “translate,” “summarize,” “responsible,” “explainable,” and “private data.” If an option is powerful but doesn’t directly satisfy the requirement, it’s a distractor.

Another trap is mixing service families that sound similar. “Azure Machine Learning” (training and managing ML) is not the same as using “Azure AI services” (prebuilt vision/language/speech capabilities). “Azure AI Language” (classical NLP tasks) is not the same as “Azure OpenAI” (LLM-driven generation and chat). The exam often tests whether you can keep these boundaries clear.

Exam Tip: Use a two-pass elimination: (1) remove anything that doesn’t match the workload (vision vs language vs ML vs gen AI), then (2) remove anything that doesn’t match the constraint (prebuilt vs custom, batch vs real-time, extract vs classify).

Watch for “scope creep” traps where a complex-sounding solution is offered for a simple task. Fundamentals questions usually want the most straightforward service that meets the need, not a multi-component architecture. Also beware of negative wording (“NOT,” “least likely,” “which is false”). Slow down on those and rephrase them into a positive statement before answering.

Finally, responsible AI content can appear as a scenario “add-on”: fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. Distractors often swap these terms; pick the principle that matches the harm described (e.g., demographic bias → fairness; inability to understand model decisions → transparency; misuse of personal data → privacy and security).

Section 1.6: Baseline diagnostic and personal weak-spot map

Your first actionable deliverable is a baseline diagnostic that produces a personal weak-spot map. Take a timed simulation under realistic conditions and capture two outputs: (1) domain-level performance (which objective areas are lowest) and (2) error-type performance (concept gaps vs misreads vs overthinking). This is how you “set up your practice environment and baseline assessment” so future study time is targeted.

Build your weak-spot map as a simple table or checklist with these columns: objective area, subtopic, confidence (high/medium/low), common trap, and next action. For example: “Computer vision → OCR: low confidence; trap: confusing Read vs general image analysis; action: review OCR use cases + do a timed set focused on text extraction scenarios.” Do the same for ML fundamentals (supervised vs unsupervised vs reinforcement; classification vs regression; training vs inference) and for generative AI (prompting basics, embeddings concept, responsible use patterns).

Exam Tip: Don’t label a domain “strong” just because you got questions right. If you guessed, mark it as weak. The exam punishes fragile knowledge because wording variations break guessing strategies.

Your study schedule should now be realistic and data-driven: allocate the majority of time to the bottom two domains while maintaining light rotation through the others to prevent decay. End each week with a timed sim to validate that weak-spot repairs are transferring into performance. If your score stagnates, check whether the issue is content or execution (timing and reading). This weak-spot map becomes the backbone of the entire marathon.

Chapter milestones
  • Understand AI-900 format, timing, and question styles
  • Register for the exam and set a realistic schedule
  • Build a study plan aligned to official domains
  • Set up your practice environment and baseline assessment
Chapter quiz

1. You are planning your AI-900 preparation. Your goal is to maximize score improvement with limited study time. Which approach BEST aligns with the AI-900 exam’s emphasis and this course’s “Mock Exam Marathon” strategy?

Show answer
Correct answer: Take a baseline assessment, then focus practice on weak domains and write a rule for each missed question to reinforce service-to-task mapping
AI-900 rewards correct workload identification and mapping tasks (for example, OCR vs classification) to the right Azure service family across official domains. A baseline assessment plus targeted practice and a review loop (capturing the rule behind each miss) directly supports that decision-making focus. Memorizing API parameters is not the exam’s emphasis and delays feedback (B). Studying only one area ignores the exam’s domain coverage (C).

2. A company asks you to set expectations for AI-900. They want to know the best way to avoid losing points due to misreading questions. What should you recommend?

Show answer
Correct answer: Treat the exam as vocabulary-and-decision based: identify the workload type first, then select the matching Azure service category and verify constraints in the scenario
AI-900 scenarios typically test recognizing the AI workload (for example, OCR, entity extraction, classification, chat completion) and selecting the appropriate Azure service family, which depends on careful reading of requirements and constraints. The exam is not primarily portal-lab navigation (B). Relying on keywords while ignoring scenario details increases errors because multiple services can share similar terms (C).

3. A startup wants to build a study plan aligned with AI-900’s official objectives. Which set of study domains MOST closely matches how the exam content is organized?

Show answer
Correct answer: AI workloads and responsible AI; machine learning principles on Azure; computer vision; natural language processing; generative AI workloads
AI-900 is organized around describing AI workloads (including responsible AI) and core workload types such as machine learning, computer vision, NLP, and generative AI patterns. The other options describe domains associated with different Azure certification tracks (networking/identity/storage in B, DevOps/Kubernetes in C), not AI-900 fundamentals.

4. You are reviewing a missed practice question: “Extract printed and handwritten text from scanned invoices.” Which rule should you record to make the correct choice obvious on future AI-900 items?

Show answer
Correct answer: OCR → Azure AI Vision Read
The task is OCR (text extraction from images/documents), which maps to Azure AI Vision Read. Azure Machine Learning is for training and managing ML models, not basic OCR extraction (B). Bot Service plus Language/Orchestrator relates to conversational bot routing and intent orchestration, not invoice text extraction (C).

5. Your manager wants you to set up a practice environment and schedule for AI-900. You must reduce cram-burnout risk and ensure practice results are diagnostic. What should you do FIRST?

Show answer
Correct answer: Schedule the exam date and create a realistic study calendar, then take an initial timed baseline assessment to identify weak spots
A realistic schedule prevents last-minute cramming, and a baseline assessment provides diagnostic data so practice time targets weak domains—matching the exam’s breadth and the course’s approach. Repeated exams without review wastes attempts and does not build the decision rules the exam tests (B). Deep-diving a single service first risks neglecting other official domains and delays feedback on weaknesses (C).

Chapter 2: Describe AI Workloads (Domain 1) — Scenarios to Services

Domain 1 of AI-900 rewards a specific skill: reading a scenario and quickly classifying the workload, the AI approach, and the Azure service family that fits—without over-engineering. Most missed points happen when candidates confuse “what the business wants” (scenario) with “how it’s implemented” (service) or when they treat every problem as machine learning. This chapter builds a repeatable mapping method: identify the workload category, decide whether the solution is predictive, generative, or analytical, then select a high-level Azure AI service family. Along the way, you’ll layer in Responsible AI requirements because many questions hide the real ask behind compliance, privacy, or transparency cues.

This chapter also supports your mock-exam marathon approach: you’ll practice fast triage (“what is this asking?”), avoid common traps (“custom model” vs “prebuilt API”), and use weak-spot analytics by tagging mistakes to one of the workload categories. Exam Tip: When you miss a question, don’t just note the right answer—label the mistake as (1) workload misclassification, (2) approach confusion, (3) service-family mismatch, or (4) Responsible AI oversight. Those four labels predict your score gains better than re-reading notes.

Practice note for Recognize AI workloads and when AI is appropriate: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match workload types to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles in exam scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed domain drill set: mixed single/multiple-answer questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize AI workloads and when AI is appropriate: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match workload types to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles in exam scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed domain drill set: mixed single/multiple-answer questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize AI workloads and when AI is appropriate: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match workload types to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: AI workload categories (vision, language, speech, decision, gen AI)

Section 2.1: AI workload categories (vision, language, speech, decision, gen AI)

The exam expects you to recognize AI workloads by the data type and the output style. Start by asking: “What is the input?” Images/video → vision. Text documents/chat messages → language (NLP). Audio streams → speech. Tabular events and rules-driven outcomes → decision. Prompts that request new content (text, code, images) → generative AI.

Vision workloads include image classification, object detection, image description, and OCR (extracting text from images). On AI-900, you’re rarely asked to design a CNN; you’re asked to name the workload correctly and choose the right service family. Language workloads include sentiment analysis, key phrase extraction, entity recognition, summarization, translation, and conversational understanding. Speech includes speech-to-text, text-to-speech, and translation of spoken language. DecisionGen AI

Common trap: candidates label any “prediction” as machine learning, but some scenarios are primarily information extraction. Example: “Extract invoice totals from scanned PDFs” is not forecasting; it’s OCR + document understanding. Another trap is confusing “language” with “gen AI.” If the output is a label or extracted fields (sentiment, entities), that’s analytical NLP. If the output is a newly written paragraph, a plan, or code, that’s generative.

Exam Tip: If the scenario mentions bounding boxes, tags, captions, OCR, or “read text in an image,” anchor on Vision. If it mentions intent, entities, sentiment, or summarization of existing text, anchor on Language. If it mentions “draft,” “compose,” “rewrite,” “chat,” or “generate,” anchor on Gen AI.

Section 2.2: Predictive vs. generative vs. analytical solutions

Section 2.2: Predictive vs. generative vs. analytical solutions

Domain 1 also checks whether you can separate three solution styles that sound similar in business language. Predictive solutions estimate an unknown value or class based on patterns in historical data (classification, regression, forecasting). Analytical solutions extract structure from unstructured content (labels, entities, key phrases, topics) or find patterns (clustering, anomaly detection) without necessarily “creating” content. Generative solutions produce new content conditioned on prompts and context (text, code, images) and often rely on large language models.

Look for cues. Predictive questions often reference training data, features, labels, or “predict next month’s sales,” “estimate churn,” “detect fraud,” “approve a loan.” Analytical questions use verbs like “extract,” “identify,” “classify text,” “recognize entities,” “summarize a document,” “detect language.” Generative questions use verbs like “write,” “draft,” “create,” “respond conversationally,” “generate a product description,” “create code,” “produce an image.”

Common trap: “summarize” can be analytical NLP (extractive or abstractive). On AI-900, summarization is typically mapped to Language services (summarization) unless the scenario explicitly frames it as a chat-based content-generation app with prompting and grounded context. Another trap: “recommendations” can be predictive/decision, not generative—recommendation engines rank items; they aren’t asked to write marketing copy.

Exam Tip: When two answers seem plausible, choose the one that matches the output type. If the output is a probability/score/label → predictive. If the output is extracted structure from existing content → analytical. If the output is novel free-form content → generative.

Section 2.3: Decision support, automation, and human-in-the-loop patterns

Section 2.3: Decision support, automation, and human-in-the-loop patterns

Many scenarios are really about how AI is used operationally: to assist a human, to automate a decision, or to route uncertain cases for review. The exam will often hide this behind phrases like “help agents triage,” “reduce manual work,” “flag for review,” or “automatically approve.” Your job is to infer the pattern and its implications for risk and Responsible AI.

Decision support means AI provides a recommendation or insight while a person remains accountable (for example, a call-center agent sees sentiment and suggested responses). Automation means the AI output directly triggers an action (approve/deny, block/unblock, route shipments). Human-in-the-loop (HITL) sits between them: the system automates low-risk/high-confidence cases and escalates low-confidence or high-impact cases to humans.

Why this matters on AI-900: “fully automated” answers are often wrong when the scenario has high stakes (healthcare, finance, employment) or when it explicitly requests explainability, auditing, or review workflows. HITL also aligns with reliability and accountability: you can reduce harm by setting thresholds, monitoring drift, and requiring manual review for edge cases.

Common trap: selecting a service based only on the data type while ignoring operational constraints. Example: a content moderation scenario is not just “language analysis”—it’s a safety workflow requiring policy thresholds and escalation. Another trap: assuming the model is always right; the exam frequently rewards acknowledging uncertainty (confidence scores) and review processes.

Exam Tip: If you see “assist,” “recommend,” “provide insights,” or “support,” think decision support. If you see “automatically,” “without human intervention,” or “real-time blocking,” check for Responsible AI cues and consider HITL unless the scenario is explicitly low risk.

Section 2.4: Responsible AI fundamentals (fairness, reliability, privacy, transparency, accountability)

Section 2.4: Responsible AI fundamentals (fairness, reliability, privacy, transparency, accountability)

Responsible AI is tested as principles applied to scenarios, not as definitions to memorize. Expect questions that ask what to do (or what principle is being addressed) when stakeholders worry about bias, outages, sensitive data, or “black box” decisions.

Fairness focuses on avoiding discrimination across groups (for example, a loan model that disadvantages a protected demographic). Typical mitigations include representative data, bias evaluation, and monitoring. Reliability & safety concerns consistent performance under expected conditions and safe behavior under unexpected inputs; think monitoring, fallbacks, and testing across edge cases. Privacy & security concerns handling personal/sensitive data appropriately (data minimization, access control, encryption, retention policies). Transparency is about making users aware they are interacting with AI, explaining limitations, and providing understandable reasons for outputs when required. Accountability means humans and organizations remain responsible: governance, audit trails, and clear ownership of outcomes.

Common trap: confusing transparency with explainability. Transparency includes disclosures (“this is AI,” “these are limitations”) and clarity about data use. Explainability is often part of transparency but is specifically about reasons behind a prediction. Another trap: treating privacy as optional when the scenario includes names, addresses, health records, or voice recordings—privacy becomes the primary constraint and may change the recommended architecture (for example, limiting data sent to external systems, enforcing retention).

Exam Tip: In scenario questions, underline the “worry statement.” If the stakeholder says “biased,” map to fairness. If they say “can’t go down,” map to reliability. If they say “PII” or “regulatory,” map to privacy. If they say “users must know,” map to transparency. If they say “who is responsible,” map to accountability and governance.

Section 2.5: Azure AI service families overview (Vision, Language, Speech, Search, OpenAI)

Section 2.5: Azure AI service families overview (Vision, Language, Speech, Search, OpenAI)

AI-900 stays high level: you’re mapping workloads to service families, not designing pipelines. Use a “family first” approach before you worry about any feature name. If the input is images/video → Azure AI Vision. If the input is text → Azure AI Language. If the input is audio → Azure AI Speech. If the user needs to find relevant content across documents → Azure AI Search. If the user wants generative chat/completions/embeddings → Azure OpenAI Service.

Azure AI Vision covers image analysis, detection, and OCR-style tasks. In exam scenarios, words like “read text from a photo,” “analyze a product image,” or “detect objects” point here. Azure AI Language supports sentiment, entities, classification, summarization, and question answering patterns over text. Azure AI Speech handles speech-to-text, text-to-speech, and translation for voice experiences. Azure AI Search supports indexing and retrieval over your data; it becomes especially important when the scenario describes “search across manuals,” “find policy clauses,” or “retrieve relevant passages.” Azure OpenAI Service is used for generative patterns like chat, completions, and embeddings-based retrieval augmentation.

Common trap: choosing Azure OpenAI for any text scenario. If the scenario just needs extraction (entities, sentiment) with deterministic outputs, Language is usually the best fit. Another trap: confusing Search with Language. Search retrieves; Language extracts and classifies. In modern architectures they combine (retrieve with Search, then generate with OpenAI), but the exam often asks for the primary capability the scenario highlights.

Exam Tip: If the question is “Which service analyzes?” think Vision/Language/Speech. If the question is “Which service finds relevant documents?” think Search. If the question is “Which service generates a response or content?” think OpenAI.

Section 2.6: Exam-style practice set and rationales for Domain 1

Section 2.6: Exam-style practice set and rationales for Domain 1

This course uses timed simulations, so your performance depends on speed and pattern recognition. Domain 1 questions are typically short, and the distractors are “near misses” (right family, wrong workload; right workload, wrong approach). Your drill strategy: (1) classify workload category, (2) classify solution style (predictive/analytical/generative), (3) identify the highest-level Azure family, (4) scan for Responsible AI constraints that override a default choice.

Rationale patterns you should practice (without memorizing specific questions):

  • Vision vs Language confusion: If the content originates as an image (scans, photos, video frames), Vision is the first stop. Language may come after OCR, but the primary capability is still Vision when the key problem is “extract text from image.”
  • Language vs OpenAI confusion: If the output is structured labels (sentiment, entities, classification), Language is favored. If the output is a novel narrative response, or the scenario mentions “prompt,” “chat,” or “drafting,” OpenAI is favored.
  • Search as the hidden requirement: If the scenario stresses “find the right document/paragraph” across a corpus, Search is usually central—even if the final interface is a chatbot.
  • Responsible AI as the deciding factor: High-impact decisions and regulated data often push you toward transparency, HITL review, and privacy-first design choices. Expect distractors that ignore these cues.

Timing guidance for the marathon: do not over-read. For each item, spend the first 5–10 seconds highlighting nouns (data type) and verbs (task). Then spend 5 seconds eliminating options that don’t match the input modality (for example, Speech options for an image-only scenario). Finally, if two remain, use the output-type rule: label/score/extract vs generate. Exam Tip: If you can’t justify an option in one sentence (“This is OCR, so Vision”), it’s probably a distractor. Your justification sentence is your best defense against tricky wording.

Weak-spot repair: after each timed drill, bucket misses into (A) modality error (vision/language/speech), (B) approach error (predictive/analytical/generative), (C) service-family confusion (Search vs Language vs OpenAI), or (D) Responsible AI oversight. Then re-run only the bucket you missed most until your reaction becomes automatic—this is the fastest path to raising your Domain 1 score.

Chapter milestones
  • Recognize AI workloads and when AI is appropriate
  • Match workload types to Azure AI services at a high level
  • Apply responsible AI principles in exam scenarios
  • Timed domain drill set: mixed single/multiple-answer questions
Chapter quiz

1. A retail company wants to automatically categorize customer emails into "billing issue," "return request," or "product question" so that tickets route to the correct team. The company has thousands of labeled examples from the past year. Which AI workload is most appropriate?

Show answer
Correct answer: Text classification (natural language processing)
This is a supervised NLP scenario: labeled emails map to discrete categories, which is text classification. Object detection applies to locating items in images, not categorizing text. Anomaly detection is for identifying unusual patterns (often numeric/time-series), not routing messages into known classes.

2. A healthcare provider wants to extract key fields (patient name, date of birth, and policy number) from scanned insurance forms that may vary slightly in layout. They want a prebuilt capability rather than training a custom model. Which Azure AI service family best fits?

Show answer
Correct answer: Azure AI Document Intelligence
Document Intelligence is designed to analyze documents and extract structured data (forms, receipts, IDs), including handling layout variation. Vision is for image analysis tasks (tags, OCR, detection) but is not the most direct fit for end-to-end form field extraction scenarios. Speech is for audio-to-text and related audio workloads, not document processing.

3. A manufacturer collects temperature readings from equipment sensors and wants to detect unusual spikes that may indicate impending failure. The goal is to flag abnormal behavior rather than predict a specific numeric value. Which workload type is this?

Show answer
Correct answer: Anomaly detection
Detecting unusual spikes compared to normal operating patterns is anomaly detection. Regression predicts a continuous value (for example, predicting the exact temperature or remaining useful life). Image classification is unrelated because the data is sensor readings, not images.

4. A company wants to add a chat experience to its website that answers questions using the company’s policy documents and returns grounded responses with citations. Which Azure AI service family is the best high-level match?

Show answer
Correct answer: Azure OpenAI
A chat experience that generates answers from documents is a generative AI workload, best aligned to Azure OpenAI for LLM-based conversational responses. Vision is for image/video analysis, not text-based chat. Azure AI Language text analytics focuses on analytical NLP (sentiment, key phrases, entity recognition) and doesn’t provide generative, multi-turn chat responses by itself.

5. A bank deploys an AI model that recommends whether to approve a loan. Auditors require the bank to provide understandable reasons for decisions and to detect potential bias across demographic groups. Which Responsible AI principle is most directly addressed by these requirements?

Show answer
Correct answer: Transparency and fairness
Providing understandable reasons maps to transparency (explainability), and checking for bias across groups maps to fairness. Reliability/performance is about consistent, accurate operation under expected conditions, not explaining decisions or bias. Privacy/security concerns protecting data and access, which is important in banking but not the primary focus of the auditor’s stated requirements.

Chapter 3: Fundamental Principles of ML on Azure (Domain 2)

Domain 2 on AI-900 tests whether you can reason about machine learning (ML) at a practical, “choose-the-right-approach” level—without needing to code. Expect questions that describe a business scenario and ask you to identify the right ML type (classification vs. regression vs. clustering), the right stage (training vs. inference), or the right Azure Machine Learning capability (AutoML, pipelines, compute). This chapter is your weak-spot repair kit: we’ll anchor the vocabulary (features/labels, algorithms/models, training/inference), then build scenario recognition patterns, then lock in evaluation metrics and overfitting awareness at exam depth.

A consistent exam theme: the test often hides the answer in plain language. Words like “predict a category,” “estimate a numeric value,” “group similar items,” “detect unusual behavior,” or “run repeatedly at scale” map directly to ML types, metrics, and Azure services. Your job is to translate the story into ML fundamentals and choose the simplest correct option.

Exam Tip: If the scenario output is a number (price, demand, temperature), think regression. If the output is a label (fraud/not fraud, churn/not churn), think classification. If there is no labeled target and the task is “discover groups,” think clustering. If the task is “find unusual,” think anomaly detection.

Practice note for Master ML fundamentals: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right ML type for a given scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand evaluation metrics and overfitting at exam depth: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed simulation: ML-focused caselets and MCQs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Master ML fundamentals: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right ML type for a given scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand evaluation metrics and overfitting at exam depth: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed simulation: ML-focused caselets and MCQs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Master ML fundamentals: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right ML type for a given scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: Core ML concepts (dataset, model, algorithm, training vs. inference)

Section 3.1: Core ML concepts (dataset, model, algorithm, training vs. inference)

AI-900 expects you to use ML terms precisely. A dataset is the collection of examples you learn from; each example is described by features (inputs) and may include a label (the correct output). The algorithm is the learning procedure (for example, logistic regression or decision trees). The model is the artifact produced after training—what you deploy and use to make predictions.

Training is the process of fitting a model to data (learning patterns). Inference is using that trained model to generate predictions on new data. Exam items frequently try to blur these two by describing “the system predicts…” while also mentioning “the model learns from…” in the same paragraph—your task is to identify what phase the system is in.

Another core term is feature engineering: selecting/transforming inputs so the model can learn effectively. On AI-900, you won’t be asked to craft code, but you may be asked which columns are features vs. labels. Example: if you are predicting “house price,” the price is the label; square footage, location, bedrooms are features.

Exam Tip: When you see “historical data with known outcomes,” that implies labeled data and supervised learning (training). When you see “predict for a new customer,” that implies inference. Don’t confuse “training dataset” with “test set”—the test set is held out for evaluation, not learning.

Common trap: assuming a model is the same as an algorithm. In exam wording, “choose an algorithm” is about the technique; “deploy a model” is about the trained result. Another trap: treating “data preprocessing” as optional—Azure ML workflows typically include data preparation steps, and questions may hint that poor data quality causes poor outcomes (a real-world and exam-relevant concept).

Section 3.2: Supervised learning (classification, regression) and common use cases

Section 3.2: Supervised learning (classification, regression) and common use cases

Supervised learning means you train using labeled data—examples where the correct answer is known. AI-900 most commonly tests two supervised tasks: classification and regression. Classification predicts a category (discrete label). Regression predicts a numeric value (continuous).

Classification patterns: “approve/deny a loan,” “spam/not spam,” “which product category,” “will the customer churn,” “detect fraud.” The output is a class label, sometimes with probabilities. Multi-class classification is still classification (more than two categories). A classic exam trap is “severity level 1–5”: although represented as numbers, it’s still classification if the numbers represent categories, not quantities.

Regression patterns: “predict sales,” “estimate delivery time,” “forecast energy consumption,” “predict temperature.” If the question emphasizes “amount,” “cost,” “time,” or “quantity,” regression is usually correct.

Exam Tip: Ask: “Would averaging two outputs make sense?” If yes, it leans regression (e.g., average predicted price). If no, it leans classification (e.g., average of ‘fraud’ and ‘not fraud’ is meaningless).

Supervised learning also introduces the idea of generalization: the model must perform well on unseen data. Overfitting (memorizing training patterns) and underfitting (too simple to learn signal) are frequently referenced indirectly, such as “performs well on training data but poorly in production.” Your best response is usually “overfitting” and/or “insufficient generalization,” not “bug in deployment.”

Common trap: confusing binary classification with anomaly detection. Fraud can be framed as either; if you have labeled fraud examples, classification fits. If you mostly have “normal” behavior and want to flag unusual cases with few labels, anomaly detection fits better (covered in Section 3.3).

Section 3.3: Unsupervised learning (clustering) and anomaly detection basics

Section 3.3: Unsupervised learning (clustering) and anomaly detection basics

Unsupervised learning uses unlabeled data. The exam’s primary unsupervised concept is clustering: grouping data points by similarity without pre-defined labels. If the scenario says “segment customers into groups,” “find similar documents,” or “group devices with similar telemetry,” clustering is the immediate association.

Clustering is often used as an exploration tool: you might cluster first, then assign business meanings to clusters (e.g., “budget shoppers,” “premium buyers”). The exam may test that clustering does not require a labeled outcome column—so if the prompt explicitly states “there is no historical outcome,” that’s a strong clue.

Anomaly detection is closely related but distinct: instead of forming groups, you identify rare or unusual observations. The exam expects you to recognize anomaly use cases like “detect credit card outliers,” “spot unusual network traffic,” “find manufacturing defects,” or “alert on sensor spikes.” Many anomaly solutions learn what “normal” looks like and flag deviations—useful when anomalies are rare and labeled examples are limited.

Exam Tip: If the goal is “create segments,” choose clustering. If the goal is “flag unusual/rare,” choose anomaly detection. Watch for wording like “outliers,” “rare events,” “deviates from baseline,” “unexpected pattern.”

Common trap: treating anomaly detection as “binary classification.” They can look similar, but classification implies you have labels for both classes (normal and anomalous). Anomaly detection is often used when you have mostly normal data and want to detect unknown or evolving anomalies.

Another trap: assuming unsupervised learning means “no evaluation.” You still evaluate clustering/anomaly models, but the metrics differ. For AI-900, focus on identifying the right approach rather than advanced cluster validity indices.

Section 3.4: Model evaluation (accuracy, precision/recall, confusion matrix, RMSE) at fundamentals level

Section 3.4: Model evaluation (accuracy, precision/recall, confusion matrix, RMSE) at fundamentals level

Evaluation questions on AI-900 are about choosing the right metric for the job and interpreting basic outcomes. For classification, the foundation is the confusion matrix, which counts true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN). From that, you get accuracy, precision, and recall.

Accuracy is the fraction of correct predictions overall. It is tempting—and often wrong—when classes are imbalanced (e.g., fraud is rare). A model can be 99% accurate by always predicting “not fraud,” but it is useless.

Precision answers: “When the model predicts positive, how often is it correct?” This matters when false positives are costly (e.g., blocking legitimate payments, sending unnecessary escalations). Recall answers: “Of all actual positives, how many did we catch?” This matters when missing a positive is costly (e.g., failing to detect fraud, missing a disease case).

Exam Tip: If the scenario emphasizes minimizing false positives, pick precision. If it emphasizes minimizing false negatives, pick recall. If it emphasizes overall correctness with balanced classes, accuracy can be acceptable.

For regression, a common metric is RMSE (root mean squared error), which measures the typical magnitude of prediction error in the same units as the target. Lower RMSE is better. If the question asks for “how close predictions are to actual numeric values,” RMSE (or related error measures) is the correct family.

Overfitting appears here too: if training performance is strong but test performance is weak, that indicates overfitting. A common exam trap is to suggest “add more features” as a remedy; often, the better conceptual answer is “use more training data,” “simplify the model,” or “regularize,” but at AI-900 level, simply recognizing overfitting/poor generalization is usually what’s tested.

Section 3.5: Azure Machine Learning basics (workspace, compute, pipelines, AutoML) and where it fits

Section 3.5: Azure Machine Learning basics (workspace, compute, pipelines, AutoML) and where it fits

Azure Machine Learning (Azure ML) is the Azure service for building, training, deploying, and managing ML models. AI-900 tests recognition of core components and when to choose Azure ML versus “prebuilt AI services.” If you need custom model training on your own data (especially for classic ML tasks), Azure ML is the go-to. If you need an out-of-the-box vision or language capability, Azure AI services may be simpler—Domain 2 focuses on Azure ML fundamentals, not the specialized APIs.

A workspace is the central container for Azure ML assets: datasets, models, experiments, endpoints, and related resources. If a question asks where you manage and organize ML resources, “workspace” is usually the answer.

Compute refers to the compute targets used for training or inference: compute instances (often for development), compute clusters (scalable training), and inference targets (endpoints). The exam often frames this as “scale training,” “run jobs on demand,” or “host a model for predictions.”

Pipelines orchestrate ML workflows as repeatable steps (data prep → train → evaluate → register → deploy). If the question emphasizes repeatability, automation, or CI/CD-like ML processes, pipelines are the concept being tested.

AutoML (automated machine learning) helps select algorithms and hyperparameters automatically. If a scenario says “we have limited ML expertise,” “quickly find the best model,” or “try multiple models automatically,” AutoML is the best fit.

Exam Tip: Watch for “custom model” versus “prebuilt.” If you must train with your own labels and choose algorithms/metrics, Azure ML is likely. If you just need OCR or sentiment analysis, the exam usually expects Azure AI services instead (covered in other domains).

Common trap: assuming Azure ML is required for all ML on Azure. The test sometimes presents Azure AI services as the simpler option. In Domain 2, however, when the story highlights model training, evaluation, deployment, and lifecycle management, Azure ML is the correct mental bucket.

Section 3.6: Exam-style practice set and rationales for Domain 2

Section 3.6: Exam-style practice set and rationales for Domain 2

This course uses timed simulations to build speed and accuracy. In Domain 2, your scoring gains come from fast scenario translation: identify (1) the ML type, (2) the phase (training vs. inference), (3) the metric, and (4) the Azure ML capability. Your timed caselets will often contain distractors that sound “more advanced” but are not required.

Rationale pattern 1—output type drives ML type: if the target is a category, it’s classification; if it’s a number, it’s regression; if there is no target and you are forming segments, it’s clustering; if the goal is “flag unusual,” it’s anomaly detection. Under time pressure, do not overthink the algorithm name; AI-900 prioritizes matching the task type.

Rationale pattern 2—metric aligns to business risk: accuracy is not the default when classes are imbalanced. Precision protects you from false positives; recall protects you from false negatives. RMSE belongs to regression. When the prompt mentions “rare events,” immediately question accuracy and look for precision/recall language cues.

Rationale pattern 3—lifecycle hints point to Azure ML features: “repeatable workflow” suggests pipelines; “limited expertise and want best model automatically” suggests AutoML; “scale training” suggests compute clusters; “central place to manage experiments and models” suggests workspace. If the scenario focuses only on consuming predictions, it’s inference and may involve a deployed endpoint.

Exam Tip: In timed sims, spend your first pass underlining (mentally) the noun phrases: “predict,” “estimate,” “group,” “outlier,” “evaluate,” “deploy,” “automate.” Those verbs usually map one-to-one to the correct option category.

Common traps in Domain 2 practice: (a) treating severity numbers as regression when they are categories; (b) picking accuracy for rare-event detection; (c) mixing up training and inference; (d) choosing Azure ML when the question is actually about a prebuilt AI service elsewhere. Your weak-spot analytics should track which trap you hit—then you drill that pattern until it becomes automatic.

Chapter milestones
  • Master ML fundamentals: features, labels, training, and inference
  • Choose the right ML type for a given scenario
  • Understand evaluation metrics and overfitting at exam depth
  • Timed simulation: ML-focused caselets and MCQs
Chapter quiz

1. A retail company wants to predict the number of units of a product it will sell next week based on price, promotions, and seasonality. Which machine learning type should you use?

Show answer
Correct answer: Regression
This is regression because the output is a numeric value (units sold). Classification is used to predict a category label (for example, high/medium/low demand), which is not requested here. Clustering is unsupervised grouping when you do not have a target value to predict.

2. You are building an Azure Machine Learning model to classify customer support tickets into categories (Billing, Technical, Account). Which items are labels in this scenario?

Show answer
Correct answer: The ticket categories (Billing, Technical, Account)
Labels are the known target values the model learns to predict—here, the ticket categories. The ticket text and metadata are features (inputs), not labels. The trained model artifact is the output of training, not the label used during training.

3. A banking app uses an existing trained model to score new loan applications submitted by customers in real time. Which stage of the ML lifecycle is occurring when the app generates the score for each new application?

Show answer
Correct answer: Inference
Generating predictions from a trained model on new data is inference. Training is the process of fitting the model using historical labeled data. Feature engineering is preparing/transforming input data, which may occur before training or inference but is not the act of scoring itself.

4. A team trained a classification model and gets 99% accuracy on the training dataset but only 70% accuracy on a held-out test dataset. What is the most likely issue?

Show answer
Correct answer: Overfitting
A large gap between training and test performance typically indicates overfitting—the model learned training-specific patterns and does not generalize. Underfitting usually shows poor performance on both training and test data. Using clustering would not produce a standard supervised accuracy comparison in this way because clustering is unsupervised and does not use labeled targets.

5. A company has historical data that includes customer demographics and whether each customer churned (Yes/No). They want the simplest Azure ML approach to automatically try algorithms and choose the best model without writing code. What should they use?

Show answer
Correct answer: Automated ML (AutoML) classification
Because the target is a labeled Yes/No outcome, this is a classification problem, and AutoML can automatically train and select a best-performing model with minimal code. Clustering is for unlabeled grouping, which does not fit a churn Yes/No target. An inference-only pipeline does not train or select a model; it assumes a trained model already exists.

Chapter 4: Computer Vision Workloads on Azure (Domain 3)

Domain 3 of AI-900 tests whether you can recognize common computer vision scenarios and map them to the right Azure service capability—quickly and confidently under time pressure. The exam is not looking for deep model-building knowledge; it is looking for correct service selection, correct feature identification (for example, “OCR” vs “image analysis”), and responsible-use awareness (especially around people and faces). This chapter follows a practical exam-coach flow: identify the vision task, map to Azure AI Vision capabilities, consider performance and responsible boundaries, then rehearse timed decision-making patterns.

In the timed sims portion of this course, vision-heavy items often hide the real requirement behind business wording like “extract text from photos of receipts” or “label the contents of product images.” Your job is to translate that wording into a workload type and output. When you do that translation first, the correct answer choices become much easier to eliminate.

Practice note for Identify vision tasks: image analysis, OCR, and detection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map tasks to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand performance considerations and responsible use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed simulation: vision-heavy items and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify vision tasks: image analysis, OCR, and detection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map tasks to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand performance considerations and responsible use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed simulation: vision-heavy items and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify vision tasks: image analysis, OCR, and detection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map tasks to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand performance considerations and responsible use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Computer vision workload patterns and typical inputs/outputs

Most AI-900 vision items reduce to three task families: (1) image analysis (describe/label what’s in an image), (2) optical character recognition (extract text), and (3) detection/recognition patterns (find things or people, sometimes with boundaries/locations). The exam commonly provides inputs such as a single image file, a set of images, a video frame, or a scanned document. The outputs, however, are what you should focus on: tags/captions, bounding boxes for objects, extracted text lines/words with coordinates, or structured fields (like invoice number) when document extraction is implied.

Start every question by asking: “Is the desired output words, labels, or locations?” Words usually means OCR. Labels/captions usually means image analysis. Locations (bounding boxes) could be object detection, people detection, or text coordinates from OCR. The exam likes to mix these signals: for example, “detect products and read the label text.” That is two tasks—image analysis plus OCR—often solvable with Azure AI Vision capabilities, but sometimes the presence of forms/invoices pushes you toward document-oriented extraction.

Exam Tip: Identify the “unit of work” and the “shape of output.” If the output is a JSON-like set of key-value fields (VendorName, Total, DueDate), that’s a document extraction pattern, not basic OCR alone.

  • Image analysis: tags, captions, objects, people counts, scenes.
  • OCR: text lines/words + coordinates; sometimes language detection.
  • Detection/recognition: bounding boxes, confidence, sometimes identity-related matches (with strong responsible-AI constraints).

Performance considerations that show up on AI-900 are typically high-level: latency (real-time vs batch), image quality, and the idea that poor lighting/angles reduce accuracy. If a scenario mentions mobile photos, glare, or skewed documents, expect OCR accuracy concerns and a recommendation to pre-process images or collect better inputs rather than assuming “a different SKU” fixes it.

Section 4.2: Image analysis concepts (tags, captions, objects, people, background)

Image analysis on Azure commonly means using Azure AI Vision to return descriptive information about an image. AI-900 expects you to recognize the difference between tags (keywords like “outdoor,” “car,” “tree”), captions (a sentence summarizing the scene), and objects (identified items, often with bounding boxes). If the scenario says “generate a description for accessibility,” that is a captioning use case. If it says “organize a photo library by categories,” tags are often the best match.

“People” and “background” cues can appear in subtle ways. “Count how many people are in the image” is still image analysis (detecting people as objects) unless the question explicitly asks to confirm identity. “Remove the background” or “identify the foreground object” is more about segmentation/background removal capabilities; the exam may phrase it as “separate a product from its background for a catalog.” You’re not expected to name model architectures—just match the scenario to a vision capability that can detect objects and/or separate background.

Common trap: confusing object detection with classification. Classification answers “what is in this image overall?” Detection answers “where are the instances of this object?” If a question includes phrases like “locate,” “bounding box,” “draw a rectangle,” or “highlight,” that is detection, not just tagging.

Exam Tip: When multiple answer choices include “Vision,” look for the verb: “analyze” (tags/captions), “detect” (objects/people and positions), “recognize text” (OCR). Verbs are often the fastest route to the correct option under timed conditions.

Responsible use also appears here: if the scenario involves people, ensure the requirement is content understanding (people present) rather than identity. AI-900 is sensitive to privacy and user consent topics when images contain people, and you should treat those as separate from generic image tagging.

Section 4.3: OCR fundamentals (printed vs. handwritten, documents vs. scenes)

OCR is the “extract text from images” workload, and AI-900 will test whether you can recognize it even when the prompt avoids the word “OCR.” Phrases like “read,” “extract,” “digitize,” “scan,” “convert to text,” and “searchable PDF” are OCR signals. Azure AI Vision includes OCR capabilities that return detected text, often with position data. The exam also distinguishes between printed versus handwritten text and between documents versus natural scenes.

Printed text (typed invoices, signs with block letters) is generally easier than handwriting. If the scenario emphasizes cursive notes, clinician handwriting, or handwritten forms, expect a mention of handwriting support and lower accuracy risks. “Documents vs scenes” is another frequent hinge: a photographed street sign is a scene; a scanned contract is a document. Both can require OCR, but documents often lead to structured extraction needs (for example, key-value pairs), while scenes focus on reading visible text reliably under varied lighting and angles.

Common trap: selecting “translation” or “language understanding” services when the prompt is purely about extracting text. Translation happens after OCR. The first step is always to get the text out.

Exam Tip: If the question mentions coordinates (line bounding boxes, word polygons) or “overlay the text position,” that strongly implies OCR output, not just “document processing.” Coordinate data helps highlight text back on the image.

Performance and quality considerations matter for OCR: skew, blur, low contrast, and compression artifacts reduce extraction quality. In practice, improving capture (lighting, focus, resolution) can outperform switching tools. On the exam, when you see constraints like “mobile camera images” and “near real-time,” pick the service that does OCR and then acknowledge that image quality and preprocessing are key performance factors.

Section 4.4: Face and identity considerations (privacy, consent, safety boundaries)

AI-900 includes responsible AI fundamentals, and in vision this often shows up as questions involving faces, people, and identity. You need to distinguish between detecting a person/facial region (a vision task) and verifying or identifying a person (an identity-sensitive task). The exam may test whether you recognize that identity scenarios require stricter privacy, consent, and governance than generic image analysis.

Key principles you should apply: obtain informed consent, minimize data collection, protect biometric data, and provide transparency about how images are used. If the scenario is employee monitoring, customer recognition, or surveillance-like tracking, expect the “responsible use” angle to matter as much as the technical choice. You may need to recommend limiting scope (for example, count people rather than identify them) or ensuring human oversight.

Common trap: assuming “face detection” automatically implies “face recognition.” Detection is about finding a face region; recognition implies matching to a known identity, which is a different and more sensitive capability. If the requirement is “blur faces” or “detect faces to crop,” that is not an identity requirement.

Exam Tip: When an answer option mentions “identify a person” or “authenticate based on face,” pause and check whether the scenario explicitly states that identity confirmation is required and that consent and compliance requirements are addressed. If not, a safer non-identity approach (detect people, not who they are) is usually the better exam choice.

Safety boundaries also include preventing misuse. Vision solutions that can be used for profiling or discrimination should be governed carefully. On AI-900, demonstrating awareness—privacy by design, bias considerations, and restricted usage—is often enough to pick the best option among close distractors.

Section 4.5: Choosing services: Azure AI Vision vs. document-oriented extraction concepts

This is the highest-yield service-selection area in Domain 3: knowing when Azure AI Vision is sufficient and when the scenario is really “document extraction.” Azure AI Vision fits image analysis (tags, captions, object detection, people detection), and OCR for images. But if the prompt describes invoices, receipts, forms, contracts, IDs, or “extract fields into a database,” the workload is drifting toward document-oriented extraction, where the goal is not just text but structured information.

Document-oriented extraction concepts include key-value pair extraction, table extraction, and consistent field mapping (for example, InvoiceDate, TotalAmount). On AI-900, you may not need to name every product SKU, but you do need to recognize that “OCR alone” produces raw text, while document extraction produces structured outputs that align to business fields. If the question emphasizes downstream automation (populate ERP fields, validate totals, route approvals), it is likely testing this distinction.

Common trap: picking Azure AI Vision OCR for an invoice scenario that explicitly asks for “supplier name, invoice number, line items.” OCR can read the words, but it does not guarantee the semantic mapping to fields without additional processing. The exam expects you to spot the implied requirement for structure.

Exam Tip: Look for nouns like “fields,” “forms,” “tables,” “line items,” “key-value,” and verbs like “populate,” “extract structured data,” or “ingest into a system.” Those are your triggers to think beyond general OCR.

Performance considerations also differ. For generic image tagging, latency and API throughput may dominate. For documents, accuracy and consistency across templates matter, and you may need to consider that different layouts require different extraction strategies. Responsible use shows up here too: documents can contain PII. Minimization, encryption, access control, and retention policies are part of the “best answer” mindset even on a fundamentals exam.

Section 4.6: Exam-style practice set and rationales for Domain 3

In timed simulations, Domain 3 questions are often “two-step”: first identify the workload (image analysis vs OCR vs detection), then identify the capability/service that matches the output. To raise your score, practice eliminating distractors by naming the output you expect. If you can say, “This must return a caption,” you can immediately discard options that focus on text extraction or language modeling.

Rationale patterns you should internalize for exam readiness: (1) If the scenario says “describe” or “categorize images,” choose image analysis. (2) If it says “read text from images,” choose OCR. (3) If it says “locate objects/people,” choose detection with bounding boxes. (4) If it says “extract fields from invoices/forms,” think document-oriented extraction rather than raw OCR.

Common trap: overfitting to keywords in answer choices instead of scenario requirements. For example, seeing “Computer Vision” and selecting it automatically—without checking whether the scenario needs structured field extraction. Always translate scenario → output → capability, then match the answer.

Exam Tip: Manage time by using a 20-second rule on first pass: decide the workload type fast, pick the best match, and flag only if two answers both fit the same output. Most wrong answers fit a different output type entirely and can be eliminated quickly.

Weak-spot repair strategy: track which confusion happens to you most—(a) tags vs objects vs captions, (b) OCR vs document extraction, or (c) people detection vs identity. Then drill those contrasts with your own one-sentence “tell.” Example tells: “Captions are sentences; tags are keywords.” “OCR gives text; documents give fields.” “Detection finds; recognition identifies.” These small heuristics are exactly what you need to stay accurate under timed conditions in vision-heavy sets.

Chapter milestones
  • Identify vision tasks: image analysis, OCR, and detection
  • Map tasks to Azure AI Vision capabilities
  • Understand performance considerations and responsible use cases
  • Timed simulation: vision-heavy items and service selection
Chapter quiz

1. A retail company wants to extract the total amount and merchant name from photos of printed receipts taken on mobile devices. Which Azure capability should you use?

Show answer
Correct answer: Optical character recognition (OCR) in Azure AI Vision
This is a text-extraction requirement (receipts with printed text), which maps to OCR capabilities in Azure AI Vision. Image classification labels what is in an image (for example, “receipt” vs “invoice”) but does not return the text content. Custom Vision object detection finds bounding boxes for known objects (for example, logos) and is not the primary service for extracting printed text like totals and merchant names.

2. You have a folder of product images and need to automatically generate tags such as "shoe," "backpack," and "watch" to improve search. You do not need to train a custom model. Which Azure AI Vision feature best fits?

Show answer
Correct answer: Image analysis (tagging) in Azure AI Vision
Auto-tagging the contents of images without custom training maps to Azure AI Vision image analysis (tags/captions). OCR is for reading text from images, not identifying objects like shoes and backpacks. Face detection is not relevant to product search and is also a sensitive scenario that requires responsible-use consideration; it would not generate general product tags.

3. A manufacturing team wants to locate and draw bounding boxes around defects on parts (for example, scratches) in images. They have labeled images of defects and will iterate as new defect types appear. Which service is the best fit?

Show answer
Correct answer: Custom Vision object detection
Bounding boxes around specific, custom defect types is an object detection workload, and the scenario includes labeled images and iteration—classic Custom Vision object detection. Prebuilt image analysis tagging is general-purpose and may not recognize domain-specific defects or return precise bounding boxes for your custom classes. OCR is unrelated because the task is not text extraction.

4. A company is building a kiosk that must verify whether a user is the same person as the ID photo they present. Which approach aligns best with AI-900 responsible AI expectations for vision scenarios?

Show answer
Correct answer: Avoid building face-based identity verification unless you have a strong justification, follow applicable laws/policies, and implement safeguards; do not treat face analysis as a generic feature-selection exercise
AI-900 emphasizes responsible use, especially for people and faces. Face-based identity verification is high-risk and must be justified and governed (policy, consent, security, bias evaluation, and compliance). OCR reading the name does not verify the person’s identity and is not a substitute for biometric verification. Image tagging cannot reliably verify identity and is not designed for authentication/verification.

5. You need to process thousands of images per minute to extract text. Latency and throughput are critical, and images may be slightly skewed or blurry. Which guidance best matches performance considerations for Azure AI Vision OCR?

Show answer
Correct answer: Use the OCR capability and optimize input quality (resolution, orientation, cropping) to improve accuracy and throughput; poor image quality increases errors and can slow processing
High-volume text extraction is still an OCR workload; performance and accuracy depend heavily on input quality (clear text, good resolution, minimal skew/blur). Object detection is not a replacement for reading text and would not produce the extracted characters/words needed. Image tagging produces descriptive labels, not the actual text content, so it cannot meet the requirement even if it returns quickly.

Chapter 5: NLP + Generative AI Workloads on Azure (Domains 4)

This chapter targets AI-900 Domains 4: natural language processing (NLP) workloads and generative AI workloads on Azure. The exam expects you to recognize common language scenarios (classification, extraction, Q&A, conversation), map them to the right Azure services, and articulate the fundamentals of generative AI (LLMs, tokens, embeddings) with responsible use patterns. You are not being tested on building production architectures; you are being tested on selecting the best-fit capability and explaining why it fits the problem statement.

As you read, keep a key exam mindset: every question is usually a matching exercise between (1) the task type (extract vs classify vs generate vs retrieve), (2) the data type (free text, chat transcripts, documents), and (3) the Azure offering (Azure AI Language, bot concepts, Azure OpenAI). Tricky wording often appears as "needs deterministic answers" (Q&A) vs "needs creative or fluent generation" (generative AI), or "must return exact entities" (extraction) vs "must categorize documents" (classification).

We will build toward the chapter lessons: selecting NLP solutions for classification, extraction, Q&A, and conversation; understanding conversational building blocks; explaining prompting and grounding; and finishing with a timed-simulation mindset for mixed NLP + generative AI objectives.

Practice note for Select NLP solutions for classification, extraction, Q&A, and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand conversational AI building blocks and orchestration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts, prompting, and grounding on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed simulation: mixed NLP + generative AI objective questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Select NLP solutions for classification, extraction, Q&A, and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand conversational AI building blocks and orchestration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts, prompting, and grounding on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed simulation: mixed NLP + generative AI objective questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Select NLP solutions for classification, extraction, Q&A, and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand conversational AI building blocks and orchestration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: NLP workload patterns (sentiment, key phrases, entities, summarization)

Section 5.1: NLP workload patterns (sentiment, key phrases, entities, summarization)

AI-900 frequently frames NLP as a set of repeatable workload patterns. Your first job is to identify the pattern from the scenario language. The most common patterns are: sentiment analysis (opinion/attitude), key phrase extraction (topic clues), named entity recognition (people/places/organizations/dates), and summarization (condense text while preserving meaning). These are typically solved with prebuilt NLP capabilities rather than custom model training at the fundamentals level.

Sentiment analysis appears when the prompt mentions customer reviews, social posts, survey comments, "positive/negative/neutral," or prioritizing angry customers. Key phrase extraction shows up when the scenario says "identify main topics" or "generate tags" from text. Entity extraction (NER/PII detection) is indicated by mentions of extracting names, addresses, invoice numbers, dates, products, or redacting personal information. Summarization appears when users want a short synopsis of a long document, call transcript, or ticket history.

Exam Tip: Watch for the verb. If the task says "extract," it is usually entity/phrase extraction. If it says "classify," it is labeling into categories. If it says "generate" or "rewrite," that is generative AI (often Azure OpenAI), not classic NLP analytics.

A classic trap is to treat summarization as the same as key phrase extraction. Key phrases are sparse keywords; summarization is coherent text. Another trap: sentiment analysis is not the same as intent detection. Intent is a conversational concept (what the user wants), typically part of a bot/NLU flow, while sentiment is emotion/opinion. On the exam, you can often eliminate wrong answers by checking whether the output is structured (entities/phrases/sentiment scores) or free-form text (summaries, rewritten content).

  • Sentiment: attitude/opinion in text.
  • Key phrases: short topical terms.
  • Entities: typed spans (person, org, location, date, etc.).
  • Summarization: concise natural-language summary.

When you see multiple requirements, prioritize the primary output. For example, "summarize and list key topics" could map to both summarization and key phrase extraction. In such cases, the exam typically expects the single best match to the main goal, unless explicitly asking for multiple features.

Section 5.2: Azure AI Language concepts (text analytics, custom text classification/extraction overview)

Section 5.2: Azure AI Language concepts (text analytics, custom text classification/extraction overview)

Azure AI Language is the umbrella service commonly associated with NLP tasks in AI-900. In exam terms, think of it as the place for text analytics features (sentiment, key phrases, entities, summarization) and for moving beyond prebuilt analysis into custom classification and custom extraction when the categories or entity types are specific to your business.

Text analytics capabilities are ideal when you can use general-purpose language understanding without training on your own labels. Scenarios: analyzing product reviews, extracting locations from news, finding key phrases in support tickets, detecting PII, or summarizing a meeting transcript. When the scenario mentions your own domain labelslike categorizing insurance claims into internal types, or extracting custom fields like "PolicyId" or "PartNumber" that arent standardyou should consider the custom options.

Custom text classification is used when the desired output is a label (or multiple labels) for a document or sentence, based on your examples. The exam may hint at this with phrases like "route tickets to the correct department," "assign a category," or "tag emails." Custom named entity recognition / extraction is used when you must pull out domain-specific fields from text, indicated by requirements such as "extract invoice IDs," "capture symptoms and medications," or "identify contract clause numbers."

Exam Tip: If the categories are stable and you can define them up front, classification fits. If the goal is to capture variable values embedded in text (IDs, names, amounts), extraction fits. A frequent trap is choosing classification when the scenario is clearly about extracting values.

At fundamentals level, you are not expected to know training pipelines, but you should recognize that custom approaches require labeled examples and iteration. Another common trap is mixing OCR with NLP: if the input is a scanned document image, you would first need OCR (computer vision) to convert it to text, and then use language features. AI-900 questions sometimes test whether you notice the input format constraint.

In weak-spot repair sessions, practice translating business phrasing into these building blocks: analyze (prebuilt), categorize (classification), extract fields (entity extraction), condense (summarization). This translation step alone improves accuracy under timed conditions.

Section 5.3: Q&A and conversational AI basics (bots, intent, utterances, dialog flow) at fundamentals level

Section 5.3: Q&A and conversational AI basics (bots, intent, utterances, dialog flow) at fundamentals level

Conversational AI questions in AI-900 usually test vocabulary and basic orchestration: bots, intents, utterances, entities (as conversational slots), and dialog flow. A bot is an application that users converse with via channels (web chat, Teams, etc.). An utterance is what the user says (a single message). An intent is the goal behind the utterance (e.g., "reset password," "track order"). A dialog or dialog flow is how the bot gathers information, confirms choices, and completes tasks across multiple turns.

Q&A workloads are different from open-ended generation. When the scenario emphasizes "answer from existing knowledge," "FAQ," "policy documents," or "provide consistent responses," you are in Q&A territory. The exam expects you to recognize that Q&A systems retrieve a relevant answer from a knowledge base rather than inventing one. This is why they are often preferred for compliance-sensitive or policy-driven responses.

Exam Tip: The fastest elimination rule: if the requirement says "must not hallucinate" or "answers must come from approved content," favor Q&A/retrieval patterns over generative-only chat.

Another tested concept is that conversational systems often require orchestration: detect intent, then route to either a Q&A response, a backend action (like creating a ticket), or a human handoff. Common traps include confusing sentiment with intent (emotion vs goal), or assuming that a bot always needs LLMs. Many bots are built with deterministic dialog flows and Q&A for reliability.

In timed questions, identify whether the conversation is task-oriented (collect info, execute action) or information-oriented (answer questions). Task-oriented flows emphasize dialog management and slot-filling; information-oriented flows emphasize Q&A retrieval. If both are present, the correct option is often the one that explicitly supports routing or integrates both patterns.

Section 5.4: Generative AI concepts (LLMs, tokens, embeddings) and when to use them

Section 5.4: Generative AI concepts (LLMs, tokens, embeddings) and when to use them

Generative AI on AI-900 is primarily conceptual: what large language models (LLMs) do, what tokens are, what embeddings represent, and when generation is appropriate. An LLM is trained to predict the next token and can produce fluent text, summarize, translate, classify, and more. Tokens are the chunks of text (not always words) that the model reads and generates; token limits matter for prompt size, conversation history, and cost. Embeddings are numeric vector representations of text (or other data) that capture semantic similarity, enabling search, clustering, and retrieval of relevant content.

Use generative AI when the output needs to be natural language that is synthesizeddrafting an email, rewriting content, summarizing in a particular style, brainstorming, or explaining code. Use embeddings-driven retrieval when the goal is to find relevant passages (semantic search) and then optionally use an LLM to compose an answer grounded in those passages.

Exam Tip: If the scenario says "find similar documents" or "semantic search" or "recommend related articles," think embeddings. If it says "write" or "compose" or "rephrase," think generation. If it says "answer using company documents," think retrieval + generation (grounding).

Common exam traps: (1) assuming LLMs are always the best choice for classification/extraction. Often, classic NLP analytics is cheaper, more predictable, and easier to validate. (2) ignoring the difference between retrieval and generation: embeddings do not generate text; they enable similarity matching. (3) overlooking token constraints: long documents may need chunking and retrieval rather than stuffing everything into the prompt.

Under time pressure, anchor on the required behavior: deterministic structured output (analytics/custom extraction) vs fluent synthesis (LLM). Then check constraints: must cite sources, must use internal data, must avoid fabrication. Those constraints push you toward grounding patterns rather than raw prompting.

Section 5.5: Azure OpenAI Service basics, prompting, and responsible generative AI

Section 5.5: Azure OpenAI Service basics, prompting, and responsible generative AI

Azure OpenAI Service provides access to OpenAI models hosted on Azure with enterprise controls. For AI-900, focus on what you do with it: chat/completions for text generation, embeddings for semantic similarity, and responsible deployment practices. When an exam question asks for "generate a draft," "summarize in a specific tone," or "create a conversational assistant," Azure OpenAI is a likely matchespecially when the desired output is open-ended natural language.

Prompting is the skill of instructing the model with clear context, constraints, and desired format. Good prompts specify role, task, inputs, output format, and guardrails (e.g., "If unknown, say you dont know"). Grounding (often via retrieval-augmented generation) reduces hallucinations by providing relevant source text for the model to use. The exam often signals grounding needs with phrases like "use company policies," "answer only from approved documents," or "include citations."

Exam Tip: When you see requirements about safety, accuracy, or approved sources, look for options mentioning grounding, retrieval, or using a data sourcenot just "use a larger model" or "increase temperature." Bigger models do not guarantee correctness.

Responsible generative AI is a high-frequency theme. You should be able to describe core concerns: harmful content, bias, privacy, security, and transparency. In scenario questions, note requests involving personal data, regulated industries, or end-user decision impacts. Your safest selection is typically the option that adds content filters, human review for high-stakes decisions, data minimization, and clear user disclosures.

Common traps include: (1) using generative AI as the sole authority for factual answers without verification, (2) forgetting that prompts and outputs can contain sensitive data, and (3) assuming that "temperature" is a safety control (it is primarily a creativity/variability control). For exam readiness, practice identifying the minimum responsible steps: restrict data, ground answers, monitor outputs, and provide a fallback path (e.g., escalate to a human) for uncertain cases.

Section 5.6: Exam-style practice set and rationales for Domains 4

Section 5.6: Exam-style practice set and rationales for Domains 4

This course is a mock-exam marathon, so your goal is not just knowledgeits speed and accuracy. For Domains 4, most missed questions come from misidentifying the workload type (analytics vs conversation vs generative) or overlooking a constraint (input is scanned, answers must be grounded, output must be structured). Use a repeatable decision checklist during timed sims.

  • Step 1: Identify the output type. Structured fields/labels  analytics or custom extraction/classification. Natural-language synthesis  generative AI.
  • Step 2: Identify the input type. If its an image/PDF scan, OCR may be required before language analysis.
  • Step 3: Identify constraints. Deterministic answers, approved sources, privacy/regulatory needs  Q&A/retrieval + grounding + responsible controls.
  • Step 4: Choose the simplest service that satisfies requirements. The exam often rewards least-complex correct mapping.

Exam Tip: Watch for distractors that are "technically possible" but not the best fit. For example, an LLM can classify text, but if the question frames it as sentiment/key phrases/entities, Azure AI Language analytics is the intended answer.

Rationale patterns you should internalize: Sentiment/key phrases/entities/summarization map to Azure AI Language text analytics features. Domain-specific labels map to custom text classification. Domain-specific fields map to custom extraction/NER. FAQ-style consistent answers map to Q&A retrieval patterns. Multi-turn task completion maps to bot dialog concepts (intent, utterances, slot filling). Generative drafting and rewriting map to Azure OpenAI chat/completions. Semantic similarity and retrieval map to embeddings; grounded answers map to retrieval + generation rather than pure prompting.

For weak-spot analytics, tag every missed item with one of four causes: (1) misread verb (extract vs classify vs generate), (2) ignored constraint (grounding/safety), (3) confused services (Language vs OpenAI vs bot/Q&A), (4) rushed elimination. Then re-run a short set focusing only on that cause. Your score improves fastest when you reduce the same mistake class repeatedly, rather than re-reading everything.

Chapter milestones
  • Select NLP solutions for classification, extraction, Q&A, and conversation
  • Understand conversational AI building blocks and orchestration
  • Explain generative AI concepts, prompting, and grounding on Azure
  • Timed simulation: mixed NLP + generative AI objective questions
Chapter quiz

1. A company wants to automatically route incoming support emails into categories such as Billing, Technical Issue, and Account Access. Which Azure capability is the best fit for this task?

Show answer
Correct answer: Azure AI Language text classification
Routing emails into predefined labels is a classification problem, which maps directly to Azure AI Language text classification. Azure OpenAI can classify via prompting, but it is not the best-fit exam answer when a deterministic NLP workload exists. Entity recognition extracts structured fields (names, dates, IDs) from text; it does not assign a document to a category.

2. You have unstructured customer messages and need to extract order numbers, delivery dates, and product names into structured fields for downstream processing. What should you use?

Show answer
Correct answer: Azure AI Language named entity recognition (NER) / extraction
Extracting specific fields (order numbers, dates, product names) is an extraction scenario, typically handled by Azure AI Language NER/entity extraction. Text classification would only label the message (for example, 'Shipping') and would not reliably return exact entities. Using an ungrounded generative model may produce plausible but incorrect values, which is risky when you need exact identifiers.

3. A team wants a chatbot that answers employee questions using content from internal policy documents. Answers must be based on the documents, and the bot should cite relevant passages. Which approach best matches this requirement?

Show answer
Correct answer: Ground an Azure OpenAI model with enterprise data retrieval (RAG) so responses are based on the policy documents
The requirement to base answers on internal documents and cite passages points to grounding via retrieval-augmented generation (RAG) with Azure OpenAI. A pure generative prompt (especially with higher temperature) can hallucinate and is not tied to the source content. Sentiment analysis measures opinion/emotion in text and does not retrieve or answer policy questions.

4. You are designing a conversational solution on Azure. Which component is primarily responsible for interpreting a user's intent from natural language input (for example, determining whether the user wants to reset a password or check an order status)?

Show answer
Correct answer: Natural language understanding (NLU) to detect intents and entities
Intent detection is an NLU task (mapping user utterances to intents and extracting entities). Speech-to-text only converts spoken audio into text; it does not infer intent. OCR extracts printed/handwritten text from images and is unrelated to interpreting conversational meaning.

5. An application uses an LLM to generate responses. You want outputs to be more deterministic and reduce variation across runs for the same prompt. Which change is most appropriate?

Show answer
Correct answer: Lower the temperature setting
Lower temperature generally makes generation more deterministic and less varied, which aligns with certification-level generative AI fundamentals. Increasing temperature increases creativity/variability and can reduce consistency. Tokens and embeddings are different concepts (tokens are units of text the model processes; embeddings are vector representations used for similarity/retrieval); swapping them does not control randomness in generation.

Chapter 6: Full Mock Exam and Final Review

This chapter is your capstone: two full mock runs (Part 1 and Part 2), a structured weak-spot analysis, and an exam-day checklist you can execute under pressure. AI-900 rewards candidates who can quickly recognize the workload type (vision vs. language vs. ML vs. generative AI), select the most appropriate Azure service or concept, and avoid “near-miss” options that sound plausible but don’t match the scenario. Your job in this chapter is not to “learn more content,” but to practice exam behavior: pacing, triage, and disciplined review.

Across both mock sets, you’ll deliberately rehearse how the exam phrases requirements (for example, “extract text from images” vs. “understand sentiment” vs. “build/train/deploy a model”), and how Responsible AI appears as constraints (fairness, transparency, privacy, reliability). You will also use weak-spot analytics to convert misses into a repeatable improvement plan. Keep your notes objective-tagged (Workloads, ML fundamentals, Vision, NLP, GenAI, Responsible AI, and exam readiness), because AI-900 grading is ultimately objective-based.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Full-length timed simulation rules and pacing plan

Section 6.1: Full-length timed simulation rules and pacing plan

Your goal in the full mock is to mirror the real test environment: timed, uninterrupted, and decision-focused. Set a single block of time, remove reference materials, and commit to finishing even if you feel uncertain. The AI-900 exam tends to mix short conceptual items with scenario-based items; the trap is spending too long on a single scenario when a later question is easier and worth the same credit.

Use a three-pass pacing plan. Pass 1: answer everything you can immediately (your “fast wins”). Pass 2: return to flagged items and read more carefully for constraints like “must minimize labeling,” “real-time,” “no-code,” or “extract printed and handwritten text.” Pass 3: make final calls and eliminate remaining options decisively.

Exam Tip: Treat every question as an “identify the workload” exercise first. Ask: Is this prediction/classification/regression (ML), image analysis/OCR (Vision), text understanding/sentiment/key phrases (Language), conversation/QnA (Bot), or content generation (Azure OpenAI)? If you can label the workload in 5–10 seconds, you cut your time per question dramatically.

  • Time boxes: Set a target average pace per question and a strict max time for any single item before flagging.
  • Flagging rule: Flag if you cannot eliminate at least two options quickly, or if the scenario includes multiple requirements.
  • Read the verb: “Detect,” “extract,” “classify,” “translate,” “generate,” “summarize,” “train,” and “deploy” point to different services and concepts.

Common pacing trap: re-reading the scenario repeatedly without extracting the requirement. Instead, underline (mentally) the deliverable (what output is needed) and the constraint (latency, data type, responsible AI requirement, skill level). That is the minimum information needed to select the correct answer.

Section 6.2: Mock exam set A (domain-balanced) with objectives mapping

Section 6.2: Mock exam set A (domain-balanced) with objectives mapping

Mock Exam Part 1 (Set A) is domain-balanced: you should see a representative spread across AI workloads, ML fundamentals, Vision, NLP, and Generative AI on Azure, plus Responsible AI. While taking Set A, annotate each item with an objective tag as you go (or immediately after). Your objective is to verify breadth: you are checking whether you can correctly classify scenarios into the right domain quickly.

Focus areas that commonly separate passing from failing in a balanced set are service selection and ML vocabulary. For example, the exam expects you to distinguish “training a custom model” (Azure Machine Learning) from “using prebuilt AI capabilities via API” (Azure AI services such as Vision or Language). Another trap is confusing Computer Vision “image analysis” (tags, captions, objects) with OCR “read text,” or mixing Language tasks (sentiment, entity recognition, key phrases) with generative tasks (drafting, summarizing in a conversational style).

Exam Tip: For each scenario, ask “Is the model prebuilt or custom?” If the scenario implies you need to build from labeled data, tune features, or manage experiments, that leans toward Azure Machine Learning. If it implies calling an endpoint for common tasks (OCR, sentiment, translation), that leans toward Azure AI services.

  • Workloads objective mapping: Identify whether the problem is classification, regression, clustering, anomaly detection, or recommendation—and whether it needs supervised vs. unsupervised learning.
  • Azure ML fundamentals mapping: Know the difference between training, validation, and testing; and concepts like features, labels, overfitting, and model evaluation metrics.
  • Vision mapping: Choose image analysis vs. OCR vs. custom vision training based on whether you need general insights, text extraction, or bespoke object categories.
  • NLP mapping: Map tasks like sentiment analysis, entity recognition, language detection, and summarization to Azure AI Language vs. generative models.
  • Responsible AI mapping: Identify mitigation steps (data review, bias checks, human oversight, privacy controls) when the scenario involves high-stakes decisions.

Set A is also where you validate your “first instinct accuracy.” If your first pass score is low, your issue is likely domain identification or service mapping; if your second pass improves significantly, your issue is reading precision and trap avoidance.

Section 6.3: Mock exam set B (hard-mode wording) with objectives mapping

Section 6.3: Mock exam set B (hard-mode wording) with objectives mapping

Mock Exam Part 2 (Set B) is intentionally phrased in “hard-mode” exam wording: vague stakeholders, extra details, and distractor options that are technically true but not the best fit. Your task is to identify what the exam is actually testing—usually a single concept like OCR vs. image classification, supervised vs. unsupervised, or Azure OpenAI patterns vs. traditional NLP APIs.

Expect traps like these: (1) “analyze documents” where the hidden requirement is text extraction (OCR) rather than sentiment; (2) “predict a value” where options try to pull you into classification when it’s regression; (3) “create a chatbot” where the exam wants the concept of intent recognition and orchestration (Bot + language understanding concepts) rather than pure generative responses; (4) “generate marketing copy” where a candidate incorrectly selects Language analytics instead of Azure OpenAI.

Exam Tip: In hard-mode wording, locate the output noun. If the output is “a label/category,” think classification; “a number,” think regression; “groups,” think clustering; “text extracted,” think OCR; “a response in natural language,” think generative AI or conversational patterns.

  • Objective mapping for GenAI: Recognize prompts, completions, embeddings, grounding, and safety filters as the tested concepts, not deep model architecture.
  • Objective mapping for Responsible AI: When scenarios mention sensitive attributes, hiring/credit, or healthcare, the exam often expects transparency, fairness evaluation, and human-in-the-loop controls.
  • Objective mapping for service selection: Prefer the most direct managed service for the requirement; avoid “build your own model” unless customization is explicitly required.

Set B should feel harder even if the underlying content is the same. If you miss questions here, it often indicates you are reacting to familiar keywords instead of reading for deliverables and constraints. Your review should focus on why the wrong option was tempting, and what wording would have disqualified it.

Section 6.4: Review method: error log, objective tags, and redo schedule

Section 6.4: Review method: error log, objective tags, and redo schedule

Your weak spot analysis should be mechanical, not emotional. Build an error log with three columns: (1) objective tag (e.g., Vision-OCR, ML-Types, GenAI-Responsible Use), (2) failure mode, and (3) corrected rule. Failure modes are typically: misread requirement, confused service, confused ML type, or “fell for distractor.” The corrected rule must be short enough to recall during the exam.

Exam Tip: Don’t just write “study OCR.” Write a decision rule: “If the output is text from an image/PDF scan, choose OCR/Read; image analysis is for objects/tags/captions.” Decision rules translate directly into points.

  • Objective tag discipline: Every miss gets exactly one primary objective tag and optionally one secondary tag (e.g., Responsible AI).
  • Redo schedule: Redo missed items after 24 hours, then again after 72 hours, without looking at your prior explanation.
  • Trap catalog: Maintain a list of your top 5 recurring traps (e.g., mixing Language analytics with Azure OpenAI generation).

When you redo, focus on speed and certainty. If you need more than a short recall to justify the answer, your understanding is still fragile. The goal is automaticity: see the scenario, label the workload, pick the service/concept, move on.

Finally, quantify progress by objective: your target is not just a higher overall score, but fewer misses in your worst two domains. That’s the fastest way to raise your passing probability.

Section 6.5: Final domain recap: service selection cheat sheet (conceptual)

Section 6.5: Final domain recap: service selection cheat sheet (conceptual)

This recap is a conceptual cheat sheet to use during your final review. AI-900 rarely requires deep implementation detail; it expects correct matching between scenario and capability. Build a mental map from “what is needed” to “what Azure offers.”

  • ML on Azure: Use Azure Machine Learning when you need to train, tune, and manage custom models; know supervised (classification/regression) vs. unsupervised (clustering), and the purpose of training/validation/testing.
  • Computer Vision: Use Azure AI Vision for image analysis (objects, captions, tags) and OCR (extract printed/handwritten text). Custom vision-like needs imply training a specialized model when the categories are domain-specific.
  • NLP: Use Azure AI Language for sentiment, key phrases, entity recognition, language detection, and intent-style understanding concepts. Distinguish analytics (extract/label) from generation (compose new text).
  • Conversational AI: Think orchestration: bots handle dialog flow; language understanding concepts route intents; generative AI can be used for richer responses but needs grounding and safety controls.
  • Generative AI: Azure OpenAI Service for chat/completions, summarization, and embeddings; focus on prompt design, grounding with your data, and responsible use (content filtering, privacy, human review).
  • Responsible AI: When stakes are high, expect principles: fairness, reliability/safety, privacy/security, inclusiveness, transparency, and accountability.

Exam Tip: If two options both “can work,” choose the one that is most direct and managed for the task described. The exam favors the simplest service that satisfies the requirement without unnecessary custom building.

Common trap: interpreting “AI” as always meaning “train a model.” Many AI-900 scenarios are solved with prebuilt APIs; only select custom training when the scenario demands it (unique labels, specialized domain, or control over the model lifecycle).

Section 6.6: Exam day readiness: time management, guessing strategy, and stress control

Section 6.6: Exam day readiness: time management, guessing strategy, and stress control

On exam day, your performance is a function of pacing and composure as much as knowledge. Start with a quick scan mindset: you are not trying to prove you’re right with a dissertation—you’re matching requirements to the best-fit concept or service. Use the same three-pass strategy from your timed simulations.

Exam Tip: If you are stuck, force an elimination. Ask: “Which option is least aligned with the required output?” Removing one wrong option reduces cognitive load and improves guessing accuracy even when you remain uncertain.

  • Time management: Watch for time sinks (long scenarios). Flag early, move on, and return when you’ve banked easier points.
  • Guessing strategy: Prefer answers that match the data type (image vs. text vs. structured data) and the action verb (extract vs. classify vs. generate). Avoid overcomplicated architectures unless the question demands them.
  • Stress control: Use a reset routine between questions: one breath, identify the workload, identify the output, pick the best match, proceed.

Common exam-day trap: changing correct answers late due to anxiety. Only change an answer if you can state a clear rule that contradicts your original choice (for example, you realized it’s OCR because the output is extracted text, not image labels). If you can’t articulate the rule, keep your first selection.

Finish with a final checklist: confirm you answered every question, revisit only flagged items, and ensure you didn’t misread “best” vs. “possible.” This chapter’s full mock runs and weak-spot repair process are designed to make exam day feel familiar—and therefore manageable.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
Chapter quiz

1. A retail company wants to build a solution that can read product labels from photos taken in a warehouse and return the extracted text to an inventory system. Which Azure AI service should you use?

Show answer
Correct answer: Azure AI Vision (OCR/Read)
Azure AI Vision provides OCR capabilities (Read) for extracting printed/handwritten text from images. Azure AI Language sentiment analysis is for classifying opinions in text, not extracting text from images. Azure Machine Learning could be used to build a custom model, but AI-900 expects choosing the purpose-built Vision OCR capability rather than training a custom solution for a standard requirement.

2. You are reviewing a practice exam result and notice you frequently confuse workloads such as language understanding vs. document text extraction. Which action is the best weak-spot repair approach for AI-900 based on objective-based study?

Show answer
Correct answer: Create an objective-tagged error log and drill scenarios that map keywords (for example, "extract text" vs. "analyze sentiment") to the correct Azure service
AI-900 is objective-based and scenario-driven; an objective-tagged miss log helps you identify patterns and map requirements to the correct workload/service. Memorizing names without scenario mapping increases near-miss mistakes because many services sound similar. Taking more mocks without structured review often repeats the same errors and is less effective than targeted remediation.

3. A team is building a customer support chatbot. They want the bot to generate natural-language responses and also cite the source policy documents used to create the answer. Which approach best aligns with this requirement?

Show answer
Correct answer: Use an Azure OpenAI model with retrieval-augmented generation (RAG) over the policy documents
RAG with an Azure OpenAI model enables generative responses grounded in your documents and supports citation patterns by retrieving relevant passages. Azure AI Vision addresses image analysis, not generating policy-based textual responses. Azure Machine Learning batch prediction on numeric metrics does not meet the requirement to generate conversational answers grounded in documents.

4. A bank is deploying an AI model that helps decide whether to approve loan applications. The compliance team requires the solution to provide explanations for decisions and ensure protected groups are not disproportionately denied. Which Responsible AI principles are most directly addressed?

Show answer
Correct answer: Transparency and fairness
Transparency covers explainability/understanding model decisions, and fairness addresses bias and disparate impact across groups. Performance and scalability are engineering concerns but do not directly satisfy explainability and bias requirements. Availability and disaster recovery relate to uptime and resilience, not fairness or explanation of decisions.

5. During the exam you have 12 questions remaining and 10 minutes left. Several questions are long scenario items with similar-sounding service options. What is the most appropriate exam-day strategy to maximize your score?

Show answer
Correct answer: Triage: answer short/high-confidence items first, mark time-consuming questions for review, and eliminate near-miss options by matching requirement keywords to workloads
AI-900 rewards pacing and disciplined review: triage prevents time sinks, and keyword-to-workload matching helps avoid near-miss distractors. Over-investing in one long question risks leaving multiple questions unanswered. Blind guessing without attempting elimination typically reduces accuracy compared to a quick pass plus targeted review.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.