HELP

+40 722 606 166

messenger@eduailast.com

Azure AI-900 Responsible AI Principles: Exam Domain Deep Dive

AI Certification Exam Prep — Beginner

Azure AI-900 Responsible AI Principles: Exam Domain Deep Dive

Azure AI-900 Responsible AI Principles: Exam Domain Deep Dive

Master AI-900 domains with clear lessons, practice, and a full mock exam.

Beginner ai-900 · microsoft · azure · azure-ai-fundamentals

Prepare confidently for Microsoft AI-900 (Azure AI Fundamentals)

This course is a domain-by-domain blueprint for learners preparing for the Microsoft AI-900: Azure AI Fundamentals exam. It is designed for beginners with basic IT literacy who want a structured path that mirrors the official exam objectives, explains the “why” behind the services, and builds skill through exam-style practice.

What this course covers (aligned to the official exam domains)

The AI-900 exam focuses on recognizing AI solution types and selecting the right Azure services for common scenarios. Across six chapters, you will study each official domain:

  • Describe AI workloads — identify workload patterns and match them to Azure AI capabilities.
  • Fundamental principles of ML on Azure — learn core ML concepts and how Azure Machine Learning fits into the lifecycle.
  • Computer vision workloads on Azure — select services for image analysis, OCR, and document processing scenarios.
  • NLP workloads on Azure — choose tools for text analytics, translation, speech, and conversational use cases.
  • Generative AI workloads on Azure — understand generative AI basics and the role of Azure OpenAI service.

Chapter-based structure designed for retention

Chapter 1 gets you exam-ready before you study content: exam registration options, scoring expectations, question types, and a practical study strategy. Chapters 2–5 provide deep, objective-mapped coverage of the domains, with dedicated practice sets written in an exam style to reinforce service selection and scenario reasoning. Chapter 6 then brings everything together with a full mock exam, weak-spot analysis, and a final review checklist you can use the day before (and the day of) the test.

Responsible AI principles woven into every domain

Even when the objective is “pick the right service,” AI-900 questions often test whether you can recognize risks and apply responsible decision-making. This course emphasizes the core Responsible AI principles—fairness, reliability & safety, privacy & security, inclusiveness, transparency, and accountability—so you can confidently answer scenario questions that involve trade-offs, constraints, and stakeholder impact.

Practice that matches the exam

You will encounter multiple-choice practice that mirrors the intent of AI-900: short scenarios, clear constraints, and options that look similar unless you understand the service boundaries. Each domain chapter ends with targeted practice to help you:

  • Distinguish ML vs rules-based solutions and when AI adds value
  • Choose between Azure Machine Learning, Azure AI services, and Azure OpenAI
  • Map vision, language, and speech needs to the correct service
  • Apply Responsible AI principles to reduce harm and improve trust

How to get started on Edu AI

If you’re ready to begin, you can Register free and start following the chapter plan immediately. Want to compare learning paths first? You can also browse all courses to find related Azure and AI fundamentals prep.

Outcome: exam-ready understanding, not memorization

By the end, you’ll have a clear mental model of each exam domain, a reliable service-selection approach for scenario questions, and the confidence that comes from completing a full mock exam under realistic constraints. Use this course as your study spine, then reinforce with Microsoft Learn and hands-on exploration as time permits.

What You Will Learn

  • Describe AI workloads and identify common AI solution scenarios on Azure (Describe AI workloads)
  • Explain core machine learning concepts and how Azure Machine Learning supports training and deployment (Fundamental principles of ML on Azure)
  • Choose Azure computer vision services for image analysis, OCR, detection, and document use cases (Computer vision workloads on Azure)
  • Select Azure NLP services for text analytics, translation, speech, and conversational AI scenarios (NLP workloads on Azure)
  • Describe generative AI concepts and Azure OpenAI service capabilities, prompts, and use cases (Generative AI workloads on Azure)
  • Apply Responsible AI principles—fairness, reliability & safety, privacy & security, inclusiveness, transparency, accountability—across exam scenarios (Describe AI workloads / Generative AI workloads on Azure)

Requirements

  • Basic IT literacy (files, web apps, cloud basics)
  • No prior certification experience required
  • No programming required (helpful but optional)
  • Access to a web browser to explore Azure documentation and learning portals

Chapter 1: AI-900 Exam Orientation and Study Strategy

  • How AI-900 is structured: domains, question styles, and timing
  • Registration and exam logistics: Pearson VUE, online vs test center
  • Scoring, retakes, accommodations, and exam policies
  • Study plan builder: how to use this course + official resources

Chapter 2: Describe AI Workloads + Responsible AI Principles

  • Recognize AI workload types and typical business scenarios
  • Match Azure AI services to workload requirements
  • Responsible AI principles and trade-offs in real scenarios
  • Domain drill: exam-style questions + explanations

Chapter 3: Fundamental Principles of ML on Azure

  • ML fundamentals: features, labels, training, inference, and evaluation
  • Supervised vs unsupervised learning and common algorithms at a high level
  • Azure Machine Learning: workspace, compute, data, and model lifecycle
  • Domain drill: exam-style questions + explanations

Chapter 4: Computer Vision Workloads on Azure

  • Vision workload patterns: image analysis, object detection, OCR
  • Azure AI Vision and Document Intelligence: choosing the right tool
  • Custom vs prebuilt vision models: when to use Custom Vision
  • Domain drill: exam-style questions + explanations

Chapter 5: NLP Workloads + Generative AI Workloads on Azure

  • NLP fundamentals and selecting Azure services for text and speech
  • Azure AI Language: sentiment, key phrases, entities, and QnA patterns
  • Generative AI concepts and Azure OpenAI service basics
  • Domain drill: exam-style questions + explanations

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist

Jordan McAllister

Microsoft Certified Trainer (MCT)

Jordan McAllister is a Microsoft Certified Trainer who designs beginner-friendly pathways to Microsoft certification success. He has coached learners through Azure fundamentals and AI-focused exams using objective-mapped practice and exam-day strategies.

Chapter 1: AI-900 Exam Orientation and Study Strategy

AI-900 (Azure AI Fundamentals) is designed to confirm that you can recognize common AI workloads, match them to the right Azure services, and explain foundational concepts (machine learning, computer vision, natural language processing, and generative AI) without needing to be a data scientist or a professional developer. This chapter sets your trajectory: how the exam is structured, how Microsoft frames the objectives, how to register smoothly, and—most importantly—how to think like the exam when selecting an answer.

Expect questions that reward clarity over depth: you rarely need to calculate metrics or write code, but you must understand “why this service” and “why not that service.” You’ll also see scenario-driven prompts where the correct answer hinges on Responsible AI principles (fairness, reliability & safety, privacy & security, inclusiveness, transparency, accountability). The exam’s most common trap is not technical complexity; it’s misreading the ask (classification vs. regression, OCR vs. object detection, chat vs. search) and choosing a plausible-but-wrong Azure product.

This chapter also introduces the study plan builder approach you’ll use throughout the course: anchor your learning to the official skills outline, then practice mapping scenarios to services and Responsible AI principles. Treat every topic as two checkboxes: (1) concept definition, and (2) correct Azure service selection under constraints (latency, privacy, interpretability, safety).

Practice note for How AI-900 is structured: domains, question styles, and timing: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Registration and exam logistics: Pearson VUE, online vs test center: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Scoring, retakes, accommodations, and exam policies: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Study plan builder: how to use this course + official resources: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How AI-900 is structured: domains, question styles, and timing: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Registration and exam logistics: Pearson VUE, online vs test center: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Scoring, retakes, accommodations, and exam policies: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Study plan builder: how to use this course + official resources: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How AI-900 is structured: domains, question styles, and timing: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: What AI-900 (Azure AI Fundamentals) validates

AI-900 validates foundational literacy across Azure AI workloads: you must identify common solution scenarios and describe how Azure services fit those scenarios. The exam expects you to recognize the difference between “building a model” (Azure Machine Learning) and “using a prebuilt AI capability” (Azure AI Services such as Vision, Language, Speech, and Azure OpenAI). It also validates that you can articulate core machine learning concepts—training vs. inference, features/labels, overfitting, and evaluation—at a high level, and select the right service to operationalize an AI solution.

From an exam-coaching standpoint, think in verbs. If the scenario says “train,” “experiment,” “evaluate,” or “deploy a custom model,” your default mental category is Azure Machine Learning. If it says “extract text,” “detect objects,” “analyze sentiment,” “translate speech,” or “summarize,” your default category is Azure AI Services or Azure OpenAI. If it says “chat with enterprise data,” consider Azure OpenAI plus a retrieval approach (often described as adding your data for grounded responses), but stay alert: AI-900 typically tests concepts and service fit, not implementation patterns.

Exam Tip: When two answers both “sound like AI,” choose the one that matches the workload type named or implied in the prompt: computer vision (images), NLP (text/language/speech), ML (custom predictive models), or generative AI (content creation and reasoning with prompts). The most common trap is selecting Azure Machine Learning for a task that is clearly a prebuilt API scenario (or vice versa).

Finally, AI-900 implicitly validates Responsible AI awareness. Even when not asked directly, scenarios may include sensitive attributes, safety constraints, or explainability needs that make some choices better aligned than others.

Section 1.2: Official exam domains and objective mapping

Microsoft organizes AI-900 by objective domains (the exact percentages can change), but the stable structure is: describe AI workloads and considerations, describe fundamental principles of machine learning on Azure, and describe features of computer vision, natural language processing, and generative AI workloads on Azure. Your first study task is objective mapping: connect each domain to (a) the concept you must define and (b) the Azure service you must select in scenarios.

Use a simple mapping table as you study (you can build it in notes):

  • Describe AI workloads: Identify workload type (prediction, vision, language, decision support) and align to Azure services; include Responsible AI principles in solution framing.
  • ML on Azure: Training vs. inference, supervised vs. unsupervised, evaluation basics; Azure Machine Learning concepts (workspaces, compute, model deployment) at a “what it does” level.
  • Computer vision: Image analysis vs. OCR vs. object detection vs. document understanding; know when to use Vision capabilities versus document-focused extraction.
  • NLP: Text analytics (sentiment, key phrases, entity recognition), translation, speech-to-text/text-to-speech, and conversational AI basics.
  • Generative AI: Prompting fundamentals, use cases (summarization, extraction, drafting), and Azure OpenAI service capabilities and constraints.

Objective mapping is how you avoid drifting into over-study. AI-900 does not require deep neural network architecture knowledge; it requires accurate categorization. A common trap is memorizing product names without connecting them to scenario triggers. Instead, practice reading a scenario and underlining the “signal words”: image/text/speech, custom vs. prebuilt, real-time vs. batch, compliance needs, and safety requirements.

Exam Tip: If a question asks “which service should you use,” eliminate options that are platforms when the scenario asks for an API (and eliminate API services when the scenario explicitly requires custom training). The exam rewards choosing the minimal correct service, not the most powerful or complex one.

Section 1.3: Registration steps, pricing, and ID requirements

AI-900 registration is typically done through Microsoft Learn’s certification pages, which route you to Pearson VUE for scheduling. You’ll choose between an online proctored exam (remote) or a test center. Both options are valid; your best choice depends on your environment and risk tolerance.

Online proctoring advantages include scheduling flexibility and avoiding travel; risks include strict room requirements and potential technical interruptions. Test centers reduce environmental uncertainty but may offer fewer time slots. Either way, plan registration early enough to secure your preferred date—especially during busy seasons when time slots fill quickly.

Pricing varies by region, and Microsoft occasionally offers discounts via events, challenges, or organizational programs. Confirm the current price during checkout; don’t rely on outdated numbers. If your employer is sponsoring the exam, verify voucher rules (expiration dates, rescheduling limits).

ID requirements are not negotiable. You must present acceptable government-issued identification that matches your registration name. Name mismatches are an avoidable failure mode: if your legal ID has a middle name or accent mark, align your candidate profile accordingly before scheduling.

Exam Tip: Treat logistics like a “pre-exam checklist” objective. Many candidates lose time and composure due to preventable issues (webcam placement, blocked network ports, unacceptable desk items). Do a system test for online exams, and for test centers arrive early with the correct ID—no exceptions.

Accommodations (e.g., extra time) must be requested in advance through Microsoft/Pearson VUE processes. Do not schedule first and assume you can add accommodations later; policies can require approval before booking.

Section 1.4: Scoring model, passing expectations, and retake rules

Microsoft exams use a scaled scoring model rather than a simple “percentage correct.” The passing score is commonly expressed as 700 on a 1000-point scale, but the exam may include unscored items used for question validation. This means you should focus on consistency across objectives rather than trying to game the scoring.

“Passing expectations” for AI-900 are best interpreted as: can you reliably choose the correct Azure service for a scenario and explain the foundational concept behind that choice? Many candidates who fail are close—they miss points in predictable clusters, such as confusing OCR with document processing, mixing up translation services with text analytics, or selecting Azure Machine Learning for prebuilt AI APIs.

Retake policies can change, but the common pattern is: if you fail, you can retake after a waiting period; subsequent retakes may have longer waits and require paying the exam fee again (unless a voucher covers it). Plan for one attempt, but build a contingency plan: schedule a “buffer window” so a retake doesn’t collide with work or school deadlines.

Exam Tip: Don’t interpret a failed attempt as “I need to study everything again.” Use the score report to identify which objective areas dropped your score. Then re-drill only those areas with scenario mapping practice—especially service selection and Responsible AI alignment.

Also know exam policies: you must follow non-disclosure rules (no sharing questions), and you can be removed for rule violations. For online exams, seemingly small behaviors (looking off-screen repeatedly, reading aloud) can be flagged. Build calm exam habits now to protect your score later.

Section 1.5: How to read exam questions (stems, distractors, keywords)

AI-900 questions are often short scenarios with a clear “ask,” but the distractors are engineered to be believable. Your job is to decode the stem: identify the workload type, determine whether the solution is custom or prebuilt, and apply any constraints (privacy, latency, interpretability, safety). Then select the smallest Azure capability that satisfies the requirement.

Start with the stem’s nouns and verbs. Nouns tell you the data type: images, documents, audio, text, structured tabular data. Verbs tell you the task: “classify,” “detect,” “extract,” “translate,” “summarize,” “generate.” Constraints usually appear as adjectives or business requirements: “must explain,” “must avoid bias,” “contains personal data,” “high availability,” “human review,” “safe responses.”

Common distractor patterns include:

  • Platform vs. service confusion: An option like Azure Machine Learning is tempting even when the scenario only needs an API call to analyze sentiment.
  • Adjacent capability confusion: OCR vs. object detection vs. image classification; translation vs. entity recognition; speech-to-text vs. text-to-speech.
  • Overkill answers: Choosing a complex pipeline when a single Azure AI Service meets the requirement.

Exam Tip: Use an elimination pass: cross out options that don’t match the input modality first (e.g., speech service for an image-only scenario). Then eliminate options that require custom training when the prompt never asks to train. Finally, align to the exact output required (text extraction vs. summarization vs. classification).

Timing strategy: don’t get stuck. AI-900 rewards steady progress. Mark challenging items for review if allowed, and return after you’ve secured the straightforward points. Many candidates lose more points from rushing the last third of the exam than from any single hard question.

Section 1.6: Responsible AI mindset for scenario questions

Responsible AI is not a “nice-to-have” on AI-900; it’s a decision lens embedded into service and solution scenarios. The exam expects you to recognize when a scenario triggers one or more principles: fairness, reliability & safety, privacy & security, inclusiveness, transparency, and accountability. Your answers should reflect what a responsible practitioner would do, not just what is technically possible.

Apply a simple mental checklist while reading scenarios:

  • Fairness: Are protected or sensitive attributes (age, gender, ethnicity) involved? Is there a risk of disparate impact? Look for wording about “bias,” “equal outcomes,” or “avoid discrimination.”
  • Reliability & safety: Is the system used in high-stakes contexts (health, finance, safety)? Look for “must not hallucinate,” “must be robust,” “must handle failures.”
  • Privacy & security: Are there personal identifiers, medical records, or confidential documents? Look for “PII,” “sensitive data,” “encryption,” “access control.”
  • Inclusiveness: Are there diverse users (languages, disabilities, accents)? Look for “accessibility,” “support multiple languages,” “varied lighting conditions.”
  • Transparency: Do users need to understand why a decision was made? Look for “explain,” “interpret,” “traceable.”
  • Accountability: Is there a human decision-maker, audit needs, or governance? Look for “human-in-the-loop,” “audit,” “oversight,” “approval.”

Exam Tip: When two options both solve the problem, the more responsible choice usually includes user disclosure, human review for high-stakes decisions, secure data handling, and monitoring for drift or harmful outputs. A classic trap is selecting “fully automated decision-making” when the scenario implies the need for review, appeal, or auditability.

In generative AI scenarios, Responsible AI often shows up as content safety, grounding, and transparency: disclose AI assistance, protect sensitive inputs, and design prompts and policies to reduce unsafe outputs. In ML scenarios, it appears as dataset quality, bias assessment, and monitoring. Carry this mindset into every domain: the correct answer is frequently the one that best balances capability with responsible constraints.

Chapter milestones
  • How AI-900 is structured: domains, question styles, and timing
  • Registration and exam logistics: Pearson VUE, online vs test center
  • Scoring, retakes, accommodations, and exam policies
  • Study plan builder: how to use this course + official resources
Chapter quiz

1. You are planning your AI-900 study strategy. Which approach best matches how the AI-900 exam typically evaluates candidates?

Show answer
Correct answer: Focus on defining AI workload types and selecting the correct Azure AI service for a given scenario, including Responsible AI considerations.
AI-900 is an Azure AI Fundamentals exam that emphasizes recognizing common AI workloads (e.g., vision, NLP, ML, generative AI) and mapping them to appropriate Azure services and concepts. Option A is wrong because the exam rarely requires coding or manual metric calculations. Option C is wrong because advanced model architecture design and tuning are beyond the fundamentals scope; the exam targets conceptual understanding and service selection.

2. A candidate schedules AI-900 and wants the most accurate expectation for question style and common pitfalls. Which statement best reflects the exam orientation described in the chapter?

Show answer
Correct answer: Most questions are scenario-driven, and a common trap is misreading the task (for example, choosing OCR when object detection is needed).
AI-900 commonly uses scenario-based prompts where you must identify the workload and choose the correct service; misinterpreting the ask (classification vs. regression, OCR vs. object detection, chat vs. search) is a frequent source of incorrect answers. Option A is wrong because scenarios and service selection are central, not parameter memorization. Option C is wrong because AI-900 does not typically test hands-on coding/debugging.

3. Your organization requires a proctored delivery with flexible scheduling, and a colleague prefers taking the exam from home. Which option correctly describes a valid Microsoft exam logistics path for AI-900?

Show answer
Correct answer: Schedule through Pearson VUE and choose either online proctoring or a test center appointment.
Microsoft certification exams such as AI-900 are commonly delivered via Pearson VUE, with choices that typically include online proctoring or in-person test centers. Option B is wrong because registration is not done through the Azure portal and delivery is not limited to test centers. Option C is wrong because Microsoft certification exams require identity verification and are not registered through GitHub Education.

4. During practice questions, you notice many scenarios reference fairness, transparency, privacy, and accountability. What is the best exam-aligned way to handle these prompts?

Show answer
Correct answer: Treat Responsible AI principles as decision criteria that can determine the best answer even when multiple services seem plausible.
AI-900 can include scenarios where Responsible AI principles (fairness, reliability & safety, privacy & security, inclusiveness, transparency, accountability) affect the correct choice, especially when options are otherwise plausible. Option A is wrong because principles may be implied by constraints (privacy, safety, interpretability). Option C is wrong because Responsible AI considerations apply broadly, including when using prebuilt services and deploying AI solutions.

5. You are building a study plan for AI-900 using this course and official resources. Which method best matches the chapter's recommended 'two checkboxes' approach?

Show answer
Correct answer: For each topic, learn the definition and practice mapping real scenarios to the correct Azure service under constraints such as privacy, latency, and interpretability.
The chapter recommends anchoring study to the official skills outline and treating each topic as: (1) concept definition and (2) correct Azure service selection in scenarios with constraints (e.g., privacy, safety, interpretability). Option B is wrong because SKU/region memorization is not the core of AI-900 fundamentals. Option C is wrong because AI-900 generally does not require end-to-end custom model training or extensive lab-style implementation.

Chapter 2: Describe AI Workloads + Responsible AI Principles

This chapter aligns directly to the AI-900 “Describe AI workloads” objective while weaving in the Responsible AI principles that appear across both classic AI scenarios (vision, NLP, ML) and generative AI. On the exam, you are rarely asked to build anything. You are asked to recognize the workload, pick the right Azure service family, and explain what “responsible” design choices look like in that scenario.

The most common mistake is to jump to a favorite service name (for example, “Azure Machine Learning”) before identifying the workload type. Build a habit: (1) name the workload (classification, prediction, detection, generation), (2) determine the data modality (tabular, images, text/audio), (3) choose the service level (prebuilt Azure AI service vs custom model with Azure Machine Learning), and (4) apply Responsible AI principles as constraints (privacy, transparency, safety, fairness).

As you read the sections, track the signals the exam gives you: words like “recommend,” “forecast,” “recognize,” “extract,” “summarize,” “chat,” “moderate,” “interpret,” “explain,” and “audit.” Those verbs typically map cleanly to a workload type and a service family. You’ll finish with an exam-style practice set in Section 2.6 (questions are placed in the practice set only, not in the chapter narrative).

Practice note for Recognize AI workload types and typical business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match Azure AI services to workload requirements: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Responsible AI principles and trade-offs in real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize AI workload types and typical business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match Azure AI services to workload requirements: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Responsible AI principles and trade-offs in real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize AI workload types and typical business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match Azure AI services to workload requirements: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: AI workloads: prediction, classification, detection, generation

Section 2.1: AI workloads: prediction, classification, detection, generation

AI-900 expects you to recognize the workload type from a business scenario. Four common buckets cover most exam prompts: prediction, classification, detection, and generation. Each has different outputs and evaluation approaches.

Prediction usually means forecasting a numeric value (regression) or a future outcome: “predict sales next month,” “estimate time to failure,” “forecast demand.” Expect signals like “forecast,” “estimate,” “continuous value,” or “trend.” Success metrics often include MAE/RMSE. Common trap: calling every “yes/no” prediction “forecasting.” If the output is a label (approve/deny), that’s classification, not regression.

Classification assigns discrete categories: spam/not spam, fraudulent/legitimate, sentiment positive/neutral/negative, image contains a cat/dog. The exam likes precision/recall confusion: high recall reduces false negatives; high precision reduces false positives. Exam Tip: If the scenario states “minimize missed fraud,” it’s pointing to recall; if it says “avoid flagging legitimate customers,” it’s pointing to precision.

Detection finds entities, anomalies, or objects—often with location or time context. Examples: object detection with bounding boxes, anomaly detection in sensor streams, identifying key phrases/entities in text (entity recognition is effectively a detection task). Trap: “recognize text in an image” is not object detection; it’s OCR/document analysis.

Generation produces new content: text completion, summarization, translation-style rewriting, code, or image generation. It is probabilistic and can hallucinate, so Responsible AI concerns (safety, transparency) appear heavily. Exam Tip: If the prompt says “create,” “draft,” “summarize,” or “chat,” think generation and consider guardrails and human review for high-stakes outputs.

In the field, real solutions combine these: a classifier routes requests, a detector extracts entities, and a generator drafts responses. On the exam, isolate the primary workload the question is measuring.

Section 2.2: Workload-to-service mapping (Azure AI services overview)

Section 2.2: Workload-to-service mapping (Azure AI services overview)

Once you identify the workload, AI-900 tests whether you can choose the appropriate Azure service. The biggest exam divider is “prebuilt AI service” versus “build your own model.” Prebuilt services (Azure AI services) are optimized for common scenarios and minimal ML expertise; custom training and MLOps typically point to Azure Machine Learning.

Computer vision workloads map to Azure AI Vision for image analysis and OCR-style tasks, and to Azure AI Document Intelligence for form/document extraction (invoices, receipts, structured fields). Trap: selecting “Vision” for key-value extraction from PDFs; the exam expects Document Intelligence when it’s forms, invoices, tables, and structured documents.

NLP workloads map to Azure AI Language for text analytics (sentiment, key phrases, entity recognition, classification, summarization), Azure AI Translator for translation, and Azure AI Speech for speech-to-text, text-to-speech, and speech translation. Conversational scenarios often map to Azure AI Bot Service combined with Language (for intent/entity understanding) or Azure OpenAI for generative chat. Exam Tip: If you see “transcribe calls,” that’s Speech; if you see “detect sentiment,” that’s Language; if you see “translate documents,” that’s Translator.

Generative AI scenarios map to Azure OpenAI Service for LLM-based chat/completions/embeddings and, depending on the question, content filtering and grounded responses (retrieval-augmented generation patterns). Trap: using generative AI to “extract fields reliably from invoices”—that’s a Document Intelligence scenario unless the question explicitly emphasizes creative drafting or open-ended language generation.

Custom model development and deployment patterns map to Azure Machine Learning: training, evaluation, model registry, endpoints, and pipelines. The exam commonly hints at “bring your own data,” “train a model,” “deploy to an endpoint,” or “MLOps.” If it asks for “no-code/low-code,” that can still be Azure Machine Learning (Automated ML, designer) rather than a prebuilt cognitive service.

When choices include multiple plausible services, use the constraint words: “structured forms,” “real-time transcription,” “object bounding boxes,” “predict numeric value,” “custom training,” and “minimal code.” Those are the exam’s breadcrumbs.

Section 2.3: Responsible AI principles: fairness, inclusiveness, transparency

Section 2.3: Responsible AI principles: fairness, inclusiveness, transparency

Responsible AI is not tested as philosophy; it’s tested as scenario judgment. You must match the principle to the risk in the prompt. Start with three principles that often show up together: fairness, inclusiveness, and transparency.

Fairness means the system should not create unjustified different outcomes for different groups. In exam scenarios, watch for protected attributes (age, gender, ethnicity, disability) and for proxies (ZIP code as a proxy for socioeconomic status). Common trap: assuming “remove protected attributes” automatically makes a model fair. Bias can remain through correlated features, label bias, and sampling bias. Exam Tip: If the scenario mentions “loan approvals,” “hiring,” “healthcare triage,” or “school admissions,” the intended principle is usually fairness, and the right mitigation often involves measuring disparate impact and reviewing training data representativeness—not just deleting columns.

Inclusiveness focuses on designing for diverse abilities, languages, and contexts. For example, a speech system should handle accents and speech impairments; a vision system should handle different lighting and skin tones; an app should support multiple languages and accessible UI. Trap: confusing inclusiveness with fairness. Inclusiveness is often about usability and access; fairness is about outcomes and treatment.

Transparency means people understand when AI is being used, what it can/can’t do, and (when appropriate) why it produced a result. In Azure terms, transparency shows up as explanations, documentation, model cards, and communicating uncertainty/limitations. For generative AI, transparency includes disclosing that text is AI-generated and citing sources when grounded. Exam Tip: If the prompt says “users must understand how decisions are made” or “explain to auditors,” the principle is transparency; if it says “make it available to people with disabilities,” it’s inclusiveness.

On AI-900, select answers that are practical controls: diverse data collection, bias evaluation, user disclosures, and explainability artifacts—not vague statements like “ensure the model is ethical.”

Section 2.4: Responsible AI principles: reliability & safety, privacy & security, accountability

Section 2.4: Responsible AI principles: reliability & safety, privacy & security, accountability

These principles are commonly tested through “what could go wrong in production?” scenarios. AI-900 wants you to recognize the operational guardrails, not implement them.

Reliability & safety means the system behaves as intended across conditions and fails safely. For vision and speech, this includes robustness to noise, lighting, accents, or device quality. For generative AI, it includes preventing harmful outputs and avoiding over-reliance on unverified responses. Trap: equating high accuracy with reliability. A model can be accurate on average and still unsafe at the edges (for example, failing on certain populations or rare but critical cases). Exam Tip: If the prompt mentions “monitoring,” “fallback,” “safe response,” “content filtering,” “testing across conditions,” it’s targeting reliability & safety.

Privacy & security covers protecting data and the system from misuse. AI scenarios often involve sensitive data: voice recordings, medical notes, ID documents, or customer chats. Expect exam cues like “PII,” “HIPAA,” “GDPR,” “confidential,” “least privilege,” “encryption,” “data residency.” For generative AI, also consider prompt injection and data leakage risks. The best answer usually involves minimizing data collection, access control, encryption, and clear retention policies—plus not sending sensitive data unnecessarily to services.

Accountability means humans remain responsible for outcomes, with governance, audit trails, and escalation paths. AI should support decision-making, not obscure who owns the decision. Trap: selecting “fully automate approvals to reduce cost” in high-stakes domains—this often conflicts with accountability and safety. Exam Tip: If the scenario asks “who is responsible when the model is wrong?” or “how do we audit decisions?” the principle is accountability; look for human oversight, logging, versioning, and documented responsibilities.

In real Azure implementations, these map to practices like monitoring drift, secure endpoints, role-based access, and change management. On AI-900, you only need to identify which principle is being exercised and which control best supports it.

Section 2.5: Risk, harm, bias, and human-in-the-loop approaches

Section 2.5: Risk, harm, bias, and human-in-the-loop approaches

AI-900 scenarios often imply “risk level.” Your job is to choose controls proportionate to harm. A movie recommendation error is low harm; an incorrect medication suggestion is high harm. The exam expects you to recognize that higher harm demands stronger oversight, validation, and limitations.

Risk and harm can be individual (privacy breach, unfair denial of service) or societal (reinforcing stereotypes, discriminatory patterns). In generative AI, harm includes toxic content, misinformation, and hallucinated citations. In vision, harm includes misidentification or overconfidence in detection. Common trap: treating all bias as intentional. Many issues are dataset-related (underrepresentation), label bias (historical discrimination encoded in labels), or deployment mismatch (model used in a new population).

Bias checkpoints typically occur at data collection, labeling, model training, evaluation, and post-deployment monitoring. A good exam answer mentions measuring performance across groups, not just overall accuracy. Another common trap is assuming that “more data” always fixes bias; if the data is systematically skewed, more of it can amplify problems.

Human-in-the-loop (HITL) is a recurring mitigation. Use HITL when the cost of errors is high, when edge cases are frequent, or when outputs must be verified (for example, document extraction that triggers payments, or generative summaries used for compliance). HITL can mean review queues, escalation policies, dual-approval, or “AI suggests, human decides.” Exam Tip: If the scenario includes legal/financial/medical decisions, the safest and most responsible option is usually to keep a human final approver and to log decisions for audit.

Also consider “guardrails” as risk controls: content filters, input validation, grounding to trusted sources, rate limiting, and clear user messaging about limitations. The exam often rewards the answer that combines a technical control (filtering/monitoring) with a process control (review/accountability).

Section 2.6: Practice set: Describe AI workloads (exam-style MCQ)

Section 2.6: Practice set: Describe AI workloads (exam-style MCQ)

This practice set is designed to mirror the AI-900 style: short scenarios that require you to (1) identify the workload and (2) select the best Azure service family while (3) applying Responsible AI constraints. Expect distractors that are “technically possible” but not “best fit.” That is a classic AI-900 pattern.

When you work multiple-choice items, apply a repeatable elimination process. First, underline the output type: number (prediction), label (classification), entity/region (detection), or new content (generation). Second, underline the modality: images/video, documents, text, or speech. Third, spot constraints like “custom training,” “minimal code,” “regulated,” “must explain,” or “avoid bias.” Finally, map to the most direct Azure service: Vision/Document Intelligence for image/document extraction, Language/Translator/Speech for NLP and audio, Azure OpenAI for generative tasks, and Azure Machine Learning for custom training and deployment.

Exam Tip: Beware of “Azure Machine Learning” as a tempting universal answer. If the scenario is a standard capability (OCR, sentiment analysis, translation), the exam usually expects the prebuilt Azure AI service rather than building a model from scratch.

Also watch for Responsible AI “tells” in the answer choices: anything that adds monitoring, human review, documentation, or privacy protections is often more correct than an option that only optimizes accuracy or speed. Common traps include: selecting automation where accountability is required, ignoring privacy when PII is present, and claiming transparency by “sharing source code” rather than providing understandable explanations and disclosures.

In the course practice set that follows this chapter, focus on the explanation rationales. On AI-900, your score improves fastest when you learn why distractors are wrong: wrong workload type, wrong modality, overbuilt solution, or missing Responsible AI control.

Chapter milestones
  • Recognize AI workload types and typical business scenarios
  • Match Azure AI services to workload requirements
  • Responsible AI principles and trade-offs in real scenarios
  • Domain drill: exam-style questions + explanations
Chapter quiz

1. A retail company wants to automatically categorize customer support emails into "Billing", "Shipping", and "Returns" to route tickets to the correct team. The company has several thousand labeled examples and wants a managed, prebuilt service rather than building infrastructure. Which Azure AI workload type and service best fit the requirement?

Show answer
Correct answer: Text classification using Azure AI Language (Text Analytics / Language service)
This is a natural language processing workload: classifying text into predefined categories. Azure AI Language provides managed text classification capabilities suited to labeled email/ticket routing scenarios. Azure AI Vision is for image/video modalities, not emails. Azure Machine Learning can build a custom classifier, but the scenario asks for a managed, prebuilt service approach and ML regression is the wrong workload type (regression predicts numeric values, not categories).

2. A hospital wants to generate a short summary of a long clinician note to help nurses quickly review key information. The solution must include safeguards to reduce the risk of generating inappropriate or unsafe content. Which combination best matches the workload and a Responsible AI control?

Show answer
Correct answer: Generative summarization with Azure OpenAI, using content filtering/moderation as a safety control
Summarizing long text is a generative AI/NLP summarization workload. Azure OpenAI is the Azure service family commonly used for generative text tasks, and adding content filtering/moderation is a key safety control to help prevent harmful outputs. Object detection is the wrong workload and modality for clinician notes. Anomaly detection is not summarization, and differential privacy is a privacy technique (not transparency) and does not directly address the key safety concern for generated text in this scenario.

3. A manufacturing company wants to identify defective items on a conveyor belt by locating scratches and missing components in images. The company needs bounding boxes around defects rather than only a pass/fail label. Which Azure AI workload type and service are most appropriate?

Show answer
Correct answer: Object detection using Azure AI Vision
Locating defects in images with bounding boxes is an object detection computer vision workload, which aligns with Azure AI Vision capabilities. Document Intelligence focuses on extracting text and structure from documents (forms, invoices), not detecting physical defects on products. Time-series forecasting predicts future numeric values from sequential data and does not address image-based defect localization.

4. A bank builds a model to help decide whether to approve personal loans. After deployment, auditors ask for a clear explanation of which factors most influenced individual decisions. Which Responsible AI principle is being emphasized, and what is the most appropriate response?

Show answer
Correct answer: Transparency; provide model interpretability/explanations for predictions and document decision logic
Auditors requesting understandable reasoning behind decisions maps to the Responsible AI principle of transparency (often paired with interpretability/explainability). Providing explanations of influential features and documenting decision logic addresses this. Privacy measures like encryption are important but do not satisfy the request for explainability; removing explanations makes transparency worse. Reliability and safety focuses on consistent performance and risk reduction, but increasing model size without addressing explainability does not meet the audit requirement and can introduce additional operational risk.

5. A company wants to predict next month’s energy consumption for each building using historical meter readings and weather data. The team is unsure whether to use a prebuilt Azure AI service or build a custom model. Which initial step best aligns with the recommended exam approach before selecting a service?

Show answer
Correct answer: Identify the workload as numeric prediction (forecasting/regression) on tabular/time-series data, then decide between a prebuilt service and Azure Machine Learning
The exam expects you to first recognize the workload type and data modality: forecasting energy use is a prediction/regression task typically using tabular/time-series data. After that, you choose the right service level (prebuilt vs custom), often Azure Machine Learning for custom forecasting. Picking Azure Machine Learning first skips workload identification, which is a common mistake highlighted in the domain guidance. Azure AI Vision is for image/video workloads and is not appropriate for meter readings and weather features.

Chapter 3: Fundamental Principles of ML on Azure

This chapter maps directly to the AI-900 objective “Explain core machine learning concepts and how Azure Machine Learning supports training and deployment.” On the exam, you are rarely asked to write code; you are asked to recognize the right ML approach for a scenario, choose the correct metric, and identify which Azure Machine Learning (Azure ML) component supports each step of the model lifecycle.

Expect scenario questions that mention data columns, predicted outcomes, evaluation results, or deployment needs (real-time vs batch). Your job is to translate that story into ML vocabulary: features and labels, training vs inference, supervised vs unsupervised learning, appropriate model type (regression/classification/clustering), and suitable validation and metrics.

Responsible AI principles still apply here: fairness in model outcomes, reliability and safety in deployment, privacy and security for data and endpoints, transparency and accountability through documentation and monitoring. Even when the question “looks like” pure ML, a better answer often includes governable workflows (versioning, tracking, monitoring) rather than ad-hoc experimentation.

Practice note for ML fundamentals: features, labels, training, inference, and evaluation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Supervised vs unsupervised learning and common algorithms at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure Machine Learning: workspace, compute, data, and model lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for ML fundamentals: features, labels, training, inference, and evaluation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Supervised vs unsupervised learning and common algorithms at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure Machine Learning: workspace, compute, data, and model lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for ML fundamentals: features, labels, training, inference, and evaluation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Supervised vs unsupervised learning and common algorithms at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: Core ML concepts (training vs inference; generalization)

Section 3.1: Core ML concepts (training vs inference; generalization)

AI-900 tests whether you can separate training from inference. Training is the process of learning patterns from historical data: you provide a dataset, define features, optionally a label, choose an algorithm, and fit a model. Inference is using that trained model to make predictions on new data—often via an endpoint or batch job.

Features are the input variables (for example, “square footage,” “number of bedrooms,” “zip code”). Labels are the known outputs you want to predict (for example, “house price”). If a scenario describes “we have past outcomes and want to predict future outcomes,” you are in supervised learning with labels. If it describes “we want to find natural groupings” without known outputs, you are likely in unsupervised learning.

The key exam concept is generalization: a model should perform well on new, unseen data—not just memorize training data. A model that performs extremely well on training data but poorly on new data is overfitting. A model that performs poorly on both training and new data may be underfitting (too simple, not enough signal, or not enough relevant features).

Exam Tip: When an answer option says “use training data to evaluate final performance,” treat it as suspicious. Proper evaluation emphasizes a holdout set (test set) or cross-validation to estimate generalization.

Common trap: confusing “training accuracy” with “model quality.” AI-900 typically rewards answers that mention validation/testing, data splits, and iterating on features rather than just “train longer” or “add more compute.” Another trap is mixing up “model” vs “service.” Azure ML trains and hosts models; other Azure AI services (Vision, Language, Speech) provide prebuilt models. If the scenario implies custom prediction from your own labeled data, it’s a strong indicator for Azure ML.

Section 3.2: Model types: regression, classification, clustering (when to use)

Section 3.2: Model types: regression, classification, clustering (when to use)

AI-900 expects you to identify the model type based on the output you need. Start with a simple question: “What is the predicted value shaped like?” If it’s a number on a continuum, you likely need regression. If it’s a category, you likely need classification. If there is no label and you want segments, you likely need clustering.

Regression predicts a numeric value, such as demand forecasting, temperature prediction, or estimating ride duration. On the exam, regression scenarios frequently use phrases like “predict the price,” “estimate the amount,” or “forecast next month’s sales.”

Classification predicts a discrete class. Binary classification examples include “fraud vs not fraud” or “churn vs not churn.” Multiclass classification examples include “route A/B/C,” “disease type,” or “document category.” Look for words like “label,” “category,” “class,” “yes/no,” or “approved/denied.”

Clustering groups similar items without predefined labels, such as customer segmentation or grouping devices by telemetry patterns. Exam wording often includes “discover groups,” “segment,” “similarity,” or “no labeled outcomes.”

At a high level, the exam may mention common algorithms, but it generally tests recognition rather than implementation. For example: linear regression (regression), logistic regression/decision trees (classification), and k-means (clustering). You don’t need to compute these; you need to match the scenario to the right approach.

Exam Tip: If the output is a probability of belonging to a class (for example, “probability of churn”), that is still classification. Don’t misclassify it as regression just because it outputs a number.

Common trap: confusing clustering with classification. If you have historical examples with known categories, it’s classification. If you do not have known categories and want the algorithm to create them, it’s clustering.

Section 3.3: Metrics and validation (accuracy, precision/recall, overfitting)

Section 3.3: Metrics and validation (accuracy, precision/recall, overfitting)

Metrics are where AI-900 often differentiates shallow understanding from exam-ready reasoning. You must choose metrics that align with the business risk described. The exam frequently provides a scenario where false positives and false negatives have different costs, pushing you toward precision/recall rather than plain accuracy.

Accuracy is the proportion of correct predictions. It’s easy to understand but can be misleading with imbalanced classes (for example, 99% “not fraud” predictions can yield 99% accuracy while missing all fraud). If the scenario mentions rare events (fraud, defects, illness), be cautious about accuracy.

Precision answers: “When the model predicts positive, how often is it correct?” High precision reduces false positives. This is important when false positives are expensive (for example, flagging legitimate transactions as fraud).

Recall answers: “Out of all actual positives, how many did the model catch?” High recall reduces false negatives. This matters when missing a positive is costly (for example, failing to detect a safety defect or a high-risk patient).

For regression, common evaluation includes error-based measures (such as MAE/MSE/RMSE). You won’t usually compute them, but you should recognize that “lower error is better,” and that regression is evaluated differently than classification.

Validation is how you estimate generalization. Typical patterns include splitting data into training and test sets, and using validation during tuning. Overfitting is a recurring theme: a model that performs well on training but poorly on test indicates overfitting.

Exam Tip: If the prompt describes “model performs well in the lab but poorly after deployment,” think distribution shift (new data differs from training data) and the need for monitoring, retraining, and robust validation. The best answer often includes operational steps, not only a different algorithm.

Common trap: choosing “increase model complexity” to fix poor test performance. If the model already overfits, more complexity can worsen generalization. Look for remedies like regularization, more training data, better feature engineering, or cross-validation.

Section 3.4: Azure Machine Learning components and workflows

Section 3.4: Azure Machine Learning components and workflows

Azure Machine Learning (Azure ML) is Microsoft’s platform for managing the end-to-end ML lifecycle: data preparation, training, evaluation, registration, and deployment. The exam expects you to recognize the major building blocks and the role each one plays.

The central resource is the Azure ML workspace. It acts as the organizing boundary for assets and governance: experiments/runs, datasets or data assets, models, endpoints, and related configurations. If a question asks “where do you manage models, compute, and experiments,” the workspace is the anchor concept.

Compute in Azure ML includes compute instances (often used for development) and compute clusters (scalable training). The exam tends to emphasize the idea: training can be resource-intensive and can scale out; inference may run on dedicated compute depending on latency and throughput requirements.

Data is typically stored in Azure services (for example, Azure Storage). In Azure ML, you reference and manage data via data connections and registered data assets. Practically, the exam checks whether you understand that Azure ML does not replace your storage account; it connects to it and tracks lineage.

The model lifecycle is a frequent scenario: train a model, evaluate it, register it (so it’s versioned and reusable), then deploy it. Registration is a governance-friendly step—important for accountability and reproducibility.

Exam Tip: In answer choices, prefer workflows that include tracking and versioning (workspace, registered model, repeatable pipeline). AI-900 often rewards “managed ML lifecycle” over one-off scripts because it aligns with reliability, transparency, and accountability.

Common trap: selecting an Azure AI prebuilt service when the scenario clearly needs custom training on your own labeled dataset. If the prompt mentions training a model with your company’s data, Azure ML is a likely fit.

Section 3.5: Deployment concepts (endpoints, batch vs real-time) and MLOps basics

Section 3.5: Deployment concepts (endpoints, batch vs real-time) and MLOps basics

Deployment is the bridge from “model in development” to “model delivering value.” AI-900 often frames this as choosing between real-time inference and batch inference, and recognizing the concept of an endpoint.

Real-time endpoints are used when you need low-latency predictions per request (for example, credit decisioning at checkout, fraud scoring during a transaction, or dynamic personalization). You typically deploy the model behind a web-accessible endpoint so applications can call it.

Batch inference is used when latency per individual record is less important than throughput and cost efficiency (for example, scoring all customers nightly for churn risk, or processing a backlog of forms). Batch is also common when you want to write predictions back to storage for downstream reporting.

MLOps basics appear indirectly: repeatability, automation, monitoring, and retraining. The exam may not use the term “CI/CD” heavily, but it tests the idea that models drift over time and must be monitored. Monitoring includes tracking performance, data drift, and failures; retraining closes the loop.

Exam Tip: If a question mentions “auditability” or “reproducibility,” think versioned models, tracked experiments, and controlled deployment processes. Those align with accountability and transparency—Responsible AI principles that can be embedded in ML operations.

Common traps: (1) assuming “deploy” means “export a file.” On Azure, deployment commonly means hosting as a managed endpoint. (2) choosing real-time endpoints for offline analytics jobs—batch is often the correct operational match.

Section 3.6: Practice set: Fundamental principles of ML on Azure

Section 3.6: Practice set: Fundamental principles of ML on Azure

This section is a drill guide (not a quiz) for how AI-900 questions are structured in this domain. When you see an exam scenario, apply a consistent decision tree: identify the goal, identify whether labels exist, pick the model type, select the evaluation approach, then map to Azure ML lifecycle steps (workspace → compute/data → train → evaluate → register → deploy/monitor).

Scenario pattern 1: “Predict a number.” Your strongest signal is a continuous output (price, demand, duration). Choose regression, mention error metrics, and highlight train/test validation. If the answers include “accuracy,” that’s likely a distractor unless the problem is classification.

Scenario pattern 2: “Approve/deny,” “spam/not spam,” “defect type.” That’s classification. If the prompt emphasizes risks of missing positives, lean toward recall; if it emphasizes cost of false alarms, lean toward precision. If the dataset is imbalanced (rare fraud), do not blindly pick accuracy.

Scenario pattern 3: “Group customers into segments without known categories.” That’s clustering (unsupervised). A common distractor is classification—reject it unless labeled groups already exist. Also watch for language that hints at similarity rather than prediction.

Scenario pattern 4: “Model worked last year but performance is degrading.” Recognize drift and the need for monitoring and retraining. In Azure ML terms, favor managed workflows with tracked runs and registered models, and a deployment approach that can be updated safely.

Exam Tip: Many AI-900 distractors are category errors: a metric from the wrong problem type, a service not meant for custom training, or evaluating on the training set. When two options look plausible, pick the one that improves generalization and operational governance (validation, monitoring, versioning).

Chapter milestones
  • ML fundamentals: features, labels, training, inference, and evaluation
  • Supervised vs unsupervised learning and common algorithms at a high level
  • Azure Machine Learning: workspace, compute, data, and model lifecycle
  • Domain drill: exam-style questions + explanations
Chapter quiz

1. A retail company wants to predict whether a customer will churn (Yes/No) based on columns such as TenureMonths, SupportTicketsLast90Days, and MonthlySpend. Which ML approach should you use?

Show answer
Correct answer: Supervised learning (binary classification) with Churn as the label
This is a supervised learning scenario because historical outcomes (churn Yes/No) are available and the goal is to predict a categorical outcome. Option A is correct: the input columns are features and Churn is the label. Option B is wrong because clustering is used when you do not have labeled outcomes and want to group similar records. Option C is wrong in exam terms because regression is used to predict a continuous numeric target; while models may output probabilities, the core task described (Yes/No) is classification.

2. A manufacturing team trained a model to detect defective parts. Only 1% of parts are defective. The team reports 99% accuracy on a test set. Which additional metric is MOST important to review to understand performance on the rare defective class?

Show answer
Correct answer: Recall (sensitivity) for the defective class
With severe class imbalance, accuracy can be misleading because predicting 'not defective' for everything can still yield ~99% accuracy. Recall for the defective class measures how many actual defects are correctly identified, which is critical for this scenario. R-squared and MAE are regression metrics, so they do not apply to a binary classification defect detection problem.

3. You are designing an Azure Machine Learning solution. Data scientists need a central place to manage datasets, compute targets, runs/experiments, registered models, and deployments. Which Azure ML component should you create first?

Show answer
Correct answer: An Azure Machine Learning workspace
The Azure ML workspace is the top-level resource that organizes the ML lifecycle: data assets, compute, experiments/runs, model registry, and endpoints. Azure Container Registry can store container images used for deployment but does not provide experiment tracking or model management. AKS can host real-time inference, but it is not the central management plane for Azure ML artifacts and would typically be attached to a workspace rather than created as the first step.

4. A bank wants to segment customers into groups based on spending patterns and account behavior. There is no pre-existing label indicating which segment each customer belongs to. Which approach is MOST appropriate?

Show answer
Correct answer: Unsupervised learning using clustering (for example, k-means)
Customer segmentation without known target labels is a classic unsupervised learning problem; clustering groups similar customers based on features. Multiclass classification requires labeled training data (known segment for each record), which the scenario explicitly lacks. Regression predicts a continuous numeric value and does not address grouping customers into discrete segments.

5. A company deploys a model that approves or rejects loan applications. They need to support two scoring patterns: (1) an immediate decision while the customer is on the website, and (2) nightly scoring of the entire applicant backlog. Which deployment approach best matches these requirements in Azure ML terms?

Show answer
Correct answer: Use a real-time endpoint for immediate decisions and a batch inference job/pipeline for nightly scoring
Real-time inference is designed for low-latency, per-request scoring (web/transaction scenarios). Batch inference is designed for scoring large datasets asynchronously (nightly/periodic jobs). Option B is wrong because batch does not satisfy the interactive latency requirement. Option C is wrong because while real-time endpoints can be called repeatedly, using them to score large backlogs is inefficient and does not align with the intended batch pattern; exams typically distinguish these two deployment modes.

Chapter 4: Computer Vision Workloads on Azure

This chapter maps to the AI-900 objective area Computer vision workloads on Azure and reinforces how the exam expects you to choose the right service for common vision scenarios: image analysis, object detection, OCR, and document processing. You will see the exam repeatedly test whether you can translate a vague business requirement ("analyze images") into the correct workload pattern (classification vs detection vs OCR vs document extraction) and then into the correct Azure service (Azure AI Vision vs Azure AI Document Intelligence vs Custom Vision).

As an exam coach, here is the recurring pattern: the question stem gives you a goal (identify objects, read text, extract invoice fields), sometimes with constraints (custom product catalog, forms vary by vendor, offline/edge), and your job is to pick the simplest service that meets requirements with minimal custom training. Another frequent trap: confusing OCR (read text) with document understanding (extract structured fields like invoice total, vendor name) or confusing image analysis tagging with object detection bounding boxes.

In addition, AI-900 increasingly expects you to connect vision choices to Responsible AI principles: privacy (faces, license plates), transparency (explain what the model returns), and reliability & safety (limitations, confidence scores, human review). Keep those principles in mind as you work through each section.

Practice note for Vision workload patterns: image analysis, object detection, OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure AI Vision and Document Intelligence: choosing the right tool: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Custom vs prebuilt vision models: when to use Custom Vision: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Vision workload patterns: image analysis, object detection, OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure AI Vision and Document Intelligence: choosing the right tool: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Custom vs prebuilt vision models: when to use Custom Vision: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Vision workload patterns: image analysis, object detection, OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Vision workload taxonomy and requirements gathering

Section 4.1: Vision workload taxonomy and requirements gathering

AI-900 questions often start with ambiguous language (“analyze images,” “detect items,” “process receipts”). Your scoring advantage comes from converting that ambiguity into a workload taxonomy. In practice (and on the exam), most computer vision scenarios land in one of three buckets: image analysis (describe, tag, recognize common objects), object detection (locate objects with bounding boxes), or OCR/document extraction (read text and/or extract fields from forms).

Requirements gathering is the hidden skill the exam measures. Ask yourself: Do we need to find something in the image (bounding boxes) or merely label the image (tags/categories)? Is the text free-form (signs, screenshots) or a structured document (invoice, receipt, ID form)? Is the content domain-specific (custom parts, proprietary products) requiring training? Are there operational constraints like real-time processing, edge deployment, or regulatory controls on data retention?

  • Image analysis: Return tags, captions, categories, and sometimes quality/safety signals. No custom training required for common concepts.
  • Object detection: Return object classes plus coordinates. Choose this when location matters (count items, track placement, verify PPE compliance).
  • OCR: Convert pixels to text (printed/handwritten depending on capability). Output is text + layout positions.
  • Document extraction: Go beyond OCR to extract structured fields/tables (invoice total, line items) and relationships.

Exam Tip: If the requirement includes phrases like “bounding box,” “where in the image,” “count instances,” or “highlight the item,” default to object detection, not image tagging.

Common trap: choosing Document Intelligence when the question only needs plain OCR. Another trap: selecting Custom Vision when prebuilt image analysis already covers generic objects. The exam rewards the “least complex service that meets requirements” mindset.

Section 4.2: Azure AI Vision capabilities (image analysis, spatial understanding basics)

Section 4.2: Azure AI Vision capabilities (image analysis, spatial understanding basics)

Azure AI Vision is your primary choice for general-purpose image understanding. On AI-900, you’re not expected to memorize every API name, but you are expected to recognize capability groups: image analysis (tags, captions, categories), detecting common visual elements, and reading text for basic OCR needs. When a question describes “generate a caption,” “tag photos,” “identify common objects,” or “detect if an image contains unsafe content,” Azure AI Vision is typically the correct service family.

Image analysis workload patterns include: describing an image for accessibility, tagging assets in a digital media library, flagging adult/racy/violent imagery for content moderation workflows, and basic brand/logo identification where supported. The exam may present a scenario like “auto-generate alt text” or “search images by keywords”—these are classic prebuilt analysis outcomes.

Spatial understanding basics show up as “reason about the arrangement of objects” or “extract relationships in the scene” (for example, a person standing near a car). AI-900 keeps this high-level: you should know that some vision services can return layout/region information and that more advanced spatial reasoning often requires specialized models or additional application logic.

Exam Tip: Watch for the word “prebuilt.” If the question emphasizes quick start, no training, and common concepts, Azure AI Vision is usually favored over Custom Vision.

Common trap: assuming “image analysis” implies object detection. Many prebuilt image analysis results are labels without precise location. If the stem requires coordinates or counting items, pivot to object detection (often via Custom Vision for custom objects or appropriate detection capability when explicitly offered).

Section 4.3: OCR and document extraction with Azure AI Document Intelligence

Section 4.3: OCR and document extraction with Azure AI Document Intelligence

OCR is about reading text. Document extraction is about understanding documents. Azure AI Document Intelligence (formerly Form Recognizer) is the exam’s go-to service when the scenario mentions invoices, receipts, purchase orders, tax forms, or “extract fields into a database.” The key distinction: Document Intelligence can perform OCR and interpret structure—key-value pairs, tables, selection marks, and consistent document layouts—often with prebuilt models for common document types.

On AI-900, you should be able to choose Document Intelligence when the output needs to be structured (for example: vendor name, invoice date, subtotal, total, line items). If the scenario says “scan receipts and populate an expense system,” selecting plain OCR is an under-solution; the exam expects Document Intelligence because it reduces custom parsing and handles layout variability.

  • Use OCR (general reading) when you only need raw text (signs, screenshots, labels) and do not require structured fields.
  • Use Document Intelligence when you need fields/tables or document-type-aware extraction (receipt totals, invoice line items).
  • Use custom document models when forms are proprietary or vary beyond what prebuilt models handle, but still have consistent patterns you can train on.

Exam Tip: If the stem mentions “forms,” “key-value pairs,” “tables,” “line items,” or “populate columns in a database,” the exam is steering you toward Document Intelligence, not general OCR.

Common traps include: confusing document processing with NLP services (text analytics) and choosing language services to “extract invoice totals.” On AI-900, structured extraction from scanned documents is a Document Intelligence scenario first; you may apply NLP later, but it’s not the primary service for capturing the data.

Section 4.4: Custom Vision: classification vs object detection scenarios

Section 4.4: Custom Vision: classification vs object detection scenarios

Custom Vision is tested as the answer when you must recognize domain-specific visual concepts not reliably covered by prebuilt models—company-specific products, specialized machine parts, unique defects, or a custom set of categories. The exam typically frames this as “train a model using labeled images” or “identify our custom item types.” Your next decision is whether you need classification or object detection.

Classification answers “What is in the image?” at an image level. Choose it when each image contains a primary label (for example, “product A vs product B,” “good vs defective,” “ripe vs unripe”). Some scenarios may involve multiple labels per image (multi-label classification), such as an image containing several attributes.

Object detection answers “Where are the objects and what are they?” Choose it when the image can contain multiple instances and location matters (counting items on a shelf, detecting helmets on workers, finding defects on a surface). The output includes bounding boxes, which enables downstream automation (cropping, counting, guiding a robot arm).

Exam Tip: The fastest way to pick between classification and detection is to look for verbs like “locate,” “highlight,” “count,” or “bounding box.” Those are object detection signals.

Common exam trap: selecting Custom Vision for generic content like “detect cats and dogs” when the question implies a common concept and no custom domain. Another trap: picking classification when the scenario needs to find multiple items in one image. The exam will often include subtle plural language (“items,” “multiple products,” “several defects”) to nudge you toward detection.

Section 4.5: Responsible vision: privacy, consent, and sensitive attributes

Section 4.5: Responsible vision: privacy, consent, and sensitive attributes

AI-900 includes Responsible AI principles across all domains, including computer vision. In vision scenarios, the highest-yield issues are privacy & security (faces, license plates, identity documents), fairness (performance variation across demographic groups), and transparency (what the system can and cannot infer). Even when the question is primarily technical (“which service?”), a secondary requirement may mention compliance, data minimization, or auditability.

Privacy and consent: If a scenario involves surveillance cameras, retail analytics, or employee monitoring, assume personally identifiable information (PII) risk. Best practice is to minimize data collected and stored, restrict access, encrypt data, and implement retention policies. If only counts are needed (foot traffic), store aggregates rather than raw images when possible.

Sensitive attributes and fairness: Be cautious with requirements that imply inferring age, gender, emotion, or other sensitive traits. The exam may test that you should avoid unnecessary sensitive inference, document limitations, and introduce human review for high-impact decisions. Reliability & safety also appears when you must set confidence thresholds and handle low-confidence predictions safely (for example, route to a human verifier).

Exam Tip: When you see “identify people,” “monitor employees,” “scan IDs,” or “store customer images,” look for the responsible control in the answer set: consent, access control, retention limits, encryption, and human oversight.

Common trap: treating confidence scores as guarantees. Vision models produce probabilistic outputs; responsible deployment includes thresholding, monitoring for drift (lighting/camera changes), and documenting intended use. Accountability shows up as maintaining audit logs and clearly assigning who reviews exceptions and updates the model.

Section 4.6: Practice set: Computer vision workloads on Azure

Section 4.6: Practice set: Computer vision workloads on Azure

This section prepares you for the style of AI-900 domain questions without listing full quiz items. Expect short scenario prompts that require you to: (1) identify the workload pattern, (2) choose the Azure service, and (3) sometimes add a responsible control. Your process should be consistent and fast.

Step 1—Spot the workload pattern. If the deliverable is “tags/captions,” think Azure AI Vision image analysis. If the deliverable is “coordinates/counting,” think object detection (often Custom Vision for custom objects). If the deliverable is “read text,” think OCR; if it’s “extract invoice fields/tables,” think Azure AI Document Intelligence.

Step 2—Choose prebuilt vs custom. Prebuilt is favored when the concepts are common and time-to-value matters. Custom is favored when the categories are unique to the business or accuracy is insufficient with prebuilt models.

Step 3—Attach operational constraints. Real-time vs batch can influence architecture, but AI-900 usually focuses on service selection rather than detailed scaling. However, note whether “no training data” is stated (prebuilt) or “labeled images available” is stated (custom training).

Exam Tip: When two answers both seem plausible, pick the one that requires less custom work and aligns exactly to the output type (labels vs boxes vs structured fields). The exam often includes one “too powerful/complex” option as a distractor.

Final trap to avoid: mixing up document extraction and NLP. After Document Intelligence extracts text/fields, you could use language services for sentiment or key phrase extraction—but if the core task is turning a scanned invoice into structured data, Document Intelligence remains the primary correct choice on AI-900.

Chapter milestones
  • Vision workload patterns: image analysis, object detection, OCR
  • Azure AI Vision and Document Intelligence: choosing the right tool
  • Custom vs prebuilt vision models: when to use Custom Vision
  • Domain drill: exam-style questions + explanations
Chapter quiz

1. A retail company wants to automatically determine whether each product photo contains a backpack, shoes, or a jacket. The images are centered product shots on a plain background, and the company does not need bounding boxes. They want the simplest Azure service that meets the requirement. Which approach should you choose?

Show answer
Correct answer: Use Azure AI Vision image analysis (tagging/classification) to identify the main objects in the image
This is an image analysis/classification-style requirement (identify what is in the image) without needing locations. Azure AI Vision image analysis can return tags/captions for common objects. Object detection (option B) is used when you need bounding boxes/locations, which the scenario does not require. Document Intelligence (option C) targets documents (forms/invoices/receipts) and structured field extraction, not general product photo classification.

2. A logistics company needs to detect and locate vehicles in traffic camera images so it can count them and draw rectangles around each vehicle in a dashboard. Which computer vision workload pattern and service best fit this requirement?

Show answer
Correct answer: Object detection with Azure AI Vision to return bounding boxes and confidence scores
The requirement explicitly includes locating objects and drawing rectangles, which maps to object detection. Azure AI Vision supports object detection-style outputs such as bounding boxes and confidence scores. OCR (option A) is for reading printed/handwritten text, not locating vehicles. Document Intelligence (option C) is for extracting structured data from documents like invoices or forms, not detecting vehicles in a scene.

3. A company scans invoices from many vendors. The layout varies by vendor, and the company needs to extract structured fields such as invoice number, total amount, and due date. Which Azure AI service should you use?

Show answer
Correct answer: Azure AI Document Intelligence to extract structured fields from invoices
Extracting specific fields (invoice number, total, due date) from semi-structured documents is a document understanding problem. Azure AI Document Intelligence is designed for this (including prebuilt document models and custom extraction). OCR alone (option A) only returns text; it does not reliably map text to fields when layouts vary, so you would still need significant custom logic. Custom Vision (option C) is for image classification/object detection; classifying by vendor does not meet the requirement to extract invoice fields.

4. A manufacturer wants to identify defects on a specific part (for example, scratches on a custom metal bracket) and draw bounding boxes around the defects. The defect types are unique to the manufacturer, and there is no suitable prebuilt model. Which option is most appropriate?

Show answer
Correct answer: Use Custom Vision to train a custom object detection model for the manufacturer’s defect classes
The scenario requires custom classes and object localization (bounding boxes) for domain-specific defects, which is a strong fit for Custom Vision object detection. Azure AI Vision image analysis (option A) provides general-purpose tagging/captioning and does not automatically learn a custom defect taxonomy without training. Document Intelligence (option C) is for document content extraction (forms, receipts, invoices), not visual defect detection on manufactured parts.

5. A city department uses images from street cameras to detect vehicles and occasionally captures faces and license plates. They want to align with Responsible AI principles while keeping the solution useful. Which action best addresses privacy while maintaining the computer vision workload?

Show answer
Correct answer: Apply redaction or blurring to faces/license plates and restrict access to images, using confidence scores to support human review when needed
Privacy-focused mitigations include minimizing exposure of sensitive data (for example, redacting/blurring faces and license plates), applying access controls, and using human review for low-confidence results—supporting privacy, reliability, and safety. Storing raw images indefinitely (option A) increases privacy risk and violates data minimization practices. Disabling confidence scores (option C) reduces transparency and can harm reliability because users lose an important signal for when to escalate to human review.

Chapter 5: NLP Workloads + Generative AI Workloads on Azure

This chapter maps directly to the AI-900 domain areas that ask you to select the right Azure AI service for language, speech, translation, conversational AI, and generative AI scenarios—while consistently applying Responsible AI principles (fairness, reliability & safety, privacy & security, inclusiveness, transparency, accountability). On the exam, you are rarely asked to design an end-to-end architecture; instead, you’ll be given a business need (“analyze customer reviews,” “translate live chat,” “build a knowledge assistant,” “generate draft responses”) and must identify the best-fit service and the correct capability name.

A major test skill is taxonomy: knowing whether a task is classification (assign a label), extraction (pull structured items like entities), summarization (compress meaning), or conversation (multi-turn intent + response). The second skill is constraint reading: the question stem often contains a key phrase like “real-time,” “on-device,” “PII,” “multiple languages,” “grounded in our documents,” or “safety filters,” which points to a specific product and configuration.

Exam Tip: When two options sound plausible, look for the one that is a managed Azure AI service aligned to the described workload (Language, Speech, Translator, Azure OpenAI). AI-900 typically rewards correct service selection over deep API syntax.

Finally, Responsible AI is not a “separate topic”—it’s embedded. If the scenario includes customer data, medical content, or decision support, expect a best-answer choice that emphasizes privacy controls, transparency (disclosing AI use), human oversight, and safety mitigations.

Practice note for NLP fundamentals and selecting Azure services for text and speech: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure AI Language: sentiment, key phrases, entities, and QnA patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Generative AI concepts and Azure OpenAI service basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for NLP fundamentals and selecting Azure services for text and speech: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure AI Language: sentiment, key phrases, entities, and QnA patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Generative AI concepts and Azure OpenAI service basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain drill: exam-style questions + explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for NLP fundamentals and selecting Azure services for text and speech: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: NLP workload taxonomy: classify, extract, summarize, converse

Section 5.1: NLP workload taxonomy: classify, extract, summarize, converse

AI-900 language questions often become easy once you classify the task into one of four buckets. Classify means predicting a category or label from text (for example, sentiment polarity, topic labels, or routing emails to departments). Extract means pulling structured signals out of unstructured text (key phrases, named entities like people/organizations/locations, or detecting PII). Summarize means generating a shorter version of a document while preserving meaning (often called extractive or abstractive summarization). Converse means enabling multi-turn interaction—understanding user intent and responding, often backed by knowledge sources.

The exam tests whether you can map these workloads to the right Azure service family and feature name. For instance, sentiment analysis is classification; named entity recognition (NER) is extraction; “generate a short summary” is summarization; “build a helpdesk chat experience” is conversation. Don’t overcomplicate by introducing custom model training unless the stem explicitly mentions training your own model or using custom labels.

Common trap: Confusing “summarize” with “extract key phrases.” Key phrases are not a summary—they are a list of terms. If the ask is “two sentences capturing the main point,” you should think summarization (or generative AI summarization, depending on phrasing), not key phrase extraction.

Exam Tip: Underline verbs in the question stem. “Detect,” “identify,” “extract,” “classify,” “summarize,” and “chat/assist” are deliberate signals used by item writers to cue the correct capability.

  • Classification: sentiment, topic labels, intent routing
  • Extraction: entities, key phrases, PII
  • Summarization: condensed document or meeting recap
  • Conversation: multi-turn Q&A, assistants, bots

Responsible AI overlay: classification outcomes can introduce fairness risk (unequal error rates across languages/dialects), extraction can raise privacy concerns (PII leakage), summarization can omit critical details (reliability), and conversation can produce unsafe content (safety). Expect best answers that include guardrails and disclosure when appropriate.

Section 5.2: Azure AI Language features (Text Analytics, NER, summarization)

Section 5.2: Azure AI Language features (Text Analytics, NER, summarization)

Azure AI Language is the core managed service for many text analytics tasks on AI-900. You should recognize feature names commonly referenced in questions: sentiment analysis, key phrase extraction, named entity recognition (NER), and summarization. The exam expects you to match each feature to a scenario and to avoid mixing it up with translation or speech services.

In practice, use Language when the input is text and the output is structured insights: “Find the products mentioned in reviews” → NER/extraction. “Identify which reviews are negative” → sentiment/classification. “Give me the main points from a long complaint email” → summarization. If the stem mentions “knowledge base,” “FAQ,” or “question answering,” treat it as a QnA pattern: you’re retrieving answers from curated content rather than generating free-form text.

Common trap: Assuming generative AI is always required. Many exam scenarios are satisfied by classic NLP—especially when the question emphasizes deterministic extraction or auditable insights. If the desired output is a list of entities or a score, Language is typically a stronger fit than Azure OpenAI.

Exam Tip: Watch for “structured output” wording: “return the entities,” “extract key phrases,” “detect PII.” Those phrases point to Azure AI Language features. Conversely, “draft,” “compose,” “rewrite,” or “generate” tends to indicate generative AI.

  • Sentiment: classify attitude (positive/negative/neutral; may include confidence scores)
  • Key phrases: extract important terms (not a narrative summary)
  • NER: identify entities (people, orgs, locations, dates, etc.)
  • Summarization: shorten text while retaining meaning

Responsible AI emphasis: if text contains customer data, apply privacy and security principles—minimize data retention, control access, and consider PII detection before downstream usage. Transparency matters too: if insights are used for decisions (routing complaints, prioritizing service), the system should be explainable and auditable (accountability).

Section 5.3: Speech and translation workloads (Azure AI Speech, Translator)

Section 5.3: Speech and translation workloads (Azure AI Speech, Translator)

AI-900 frequently distinguishes between text workloads and speech workloads. When the input or output is audio, the best-fit service is typically Azure AI Speech. Core concepts you should know include speech-to-text (transcription), text-to-speech (voice synthesis), and often real-time versus batch processing. If the scenario mentions call centers, meeting transcription, captions, or voice bots, the “Speech” keyword is your strongest cue.

For language conversion between human languages, use Translator (Azure AI Translator). The exam commonly tests simple mapping: “translate a support ticket into English” → Translator; “transcribe a customer call” → Speech; “analyze the sentiment of a Spanish review” might require translation or might be supported directly depending on language coverage—but exam items usually keep it straightforward and name the needed step.

Common trap: Confusing transcription with translation. Transcription converts audio to text in the same language. Translation converts content from one language to another. If the requirement is “convert spoken French into written English,” you’re looking at a pipeline: Speech-to-text (French) + Translator (to English), unless the question provides a single best-fit choice for one step.

Exam Tip: Look for “real-time,” “captions,” “voice,” “microphone,” or “call recording.” These words reliably indicate Azure AI Speech. Look for “multi-language,” “localize,” or “convert to/from language X.” Those point to Translator.

Responsible AI concerns show up as reliability and inclusiveness: speech systems can underperform across accents, background noise, and dialects. The best practice answer often includes testing across representative user groups and providing accessibility options (for example, captions plus audio, adjustable speaking rate, or alternative input methods).

Section 5.4: Conversational AI basics (bots, orchestration concepts, safety considerations)

Section 5.4: Conversational AI basics (bots, orchestration concepts, safety considerations)

Conversational AI on AI-900 is about recognizing the components of a chatbot or assistant scenario and choosing the correct Azure approach. A typical pattern includes: a channel (web/mobile/Teams), a bot or application layer, an orchestration layer to decide what to do with user messages, and a knowledge source (FAQs, documents, or databases). Some questions describe “Q&A” behavior—answering users from a curated knowledge base—while others describe open-ended assistance, which may be generative.

From an exam perspective, the key is to separate retrieval from generation. Retrieval-based QnA returns answers grounded in known content and tends to be easier to audit (transparency, accountability). Generative chat can be more flexible but increases safety and reliability risks (hallucinations, harmful content). When the stem emphasizes “must answer only from our policy documents,” choose the approach that enforces grounding and limits free-form invention.

Common trap: Treating “bot” as a single product. The exam often expects you to recognize that conversation involves multiple services: language understanding for intent, knowledge retrieval for answers, and potentially Azure OpenAI for natural language generation. If the question is asking for “which service provides X capability,” pick the service that directly provides it, not the entire bot stack.

Exam Tip: Safety considerations are testable. If a scenario involves public-facing chat, the strongest answer typically includes content filtering, prompt/response monitoring, user disclosure that they are interacting with AI (transparency), and a human escalation path (accountability and safety).

  • Orchestration concept: route the user request to search/QnA/action/generative response based on intent and context
  • Grounding concept: constrain responses to trusted data sources to reduce hallucinations
  • Escalation: hand off to an agent for sensitive or high-risk requests

Privacy is also central: conversational logs may contain PII. The best practice choice emphasizes secure storage, access control, and data minimization. Inclusiveness: provide accessible interaction options and support diverse languages when required.

Section 5.5: Generative AI on Azure: Azure OpenAI, prompting basics, grounding concepts

Section 5.5: Generative AI on Azure: Azure OpenAI, prompting basics, grounding concepts

Generative AI refers to models that produce new content—text, summaries, code-like output, or structured responses—based on a prompt. On AI-900, your anchor service is Azure OpenAI Service. The exam expects you to identify typical use cases: drafting emails, summarizing content, extracting structured data via instructions, creating chat assistants, and improving search or helpdesk experiences.

Prompting basics are fair game. A strong prompt commonly includes: (1) role/instructions, (2) context, (3) constraints (format, tone, length), and (4) examples when needed. If a scenario says “responses must be in JSON” or “use a professional tone,” the correct concept is to specify constraints in the prompt. If it says “reduce hallucinations” or “answer only using company documents,” the correct concept is grounding—augmenting the model with trusted data (often via retrieval) and instructing it to cite or limit to that data.

Common trap: Believing generative models are deterministic and always correct. Reliability & safety is a major Responsible AI angle: models can hallucinate, be overconfident, or produce unsafe outputs. Exam answers that mention “validate outputs,” “human review,” “use grounding,” or “apply content filtering” are often the best-choice options for high-stakes scenarios.

Exam Tip: Distinguish “extract” versus “generate.” If you need consistent, auditable extraction (entities, PII), classic Language features may be preferable. If you need natural, fluent drafting or multi-turn assistance, Azure OpenAI is more likely correct.

Responsible AI mapping for generative AI is explicit: transparency (disclose AI-generated content), accountability (human-in-the-loop for impactful decisions), privacy & security (protect prompts and data sources), fairness (evaluate performance across user groups), and reliability & safety (filters, grounding, monitoring). Expect scenario questions where the “best” solution is the one that adds guardrails rather than the one that is simply the most powerful.

Section 5.6: Practice set: NLP workloads + Generative AI workloads on Azure

Section 5.6: Practice set: NLP workloads + Generative AI workloads on Azure

This section prepares you for the exam’s most common pattern: a short scenario followed by multiple plausible Azure services. Your job is to (1) classify the workload, (2) identify the data modality (text vs audio vs multilingual), (3) pick the managed service that directly matches the required capability, and (4) apply Responsible AI constraints mentioned in the stem. Many candidates miss points by choosing a tool they “like” instead of the one the prompt explicitly describes.

Use a fast decision checklist during practice: If it’s audio, start with Azure AI Speech. If it’s language-to-language conversion, start with Translator. If it’s structured insights from text (sentiment, entities, key phrases, PII), start with Azure AI Language. If it’s drafting/generating a response, summarizing conversationally, or doing flexible chat assistance, consider Azure OpenAI—then ask whether the scenario demands grounding (answers must come from company content).

Exam Tip: When you see “must not fabricate,” “only answer from our docs,” or “include citations,” that’s your signal to choose an approach that supports grounding and retrieval, not pure free-form generation.

Also practice spotting Responsible AI “best answer” cues. If the scenario includes personal data, prefer answers that mention access control, data minimization, and secure handling (privacy & security). If it’s a public chatbot, prefer answers that add content filtering, escalation, and monitoring (reliability & safety, accountability). If it serves diverse users, look for multilingual support and accessibility options (inclusiveness).

Finally, remember the exam is vocabulary-driven. Learn to match terms like “NER,” “key phrase extraction,” “speech-to-text,” “text-to-speech,” “translation,” “question answering,” “prompt,” and “grounding” to the correct service family. That mapping—plus awareness of the common traps above—is the highest ROI preparation you can do for this domain.

Chapter milestones
  • NLP fundamentals and selecting Azure services for text and speech
  • Azure AI Language: sentiment, key phrases, entities, and QnA patterns
  • Generative AI concepts and Azure OpenAI service basics
  • Domain drill: exam-style questions + explanations
Chapter quiz

1. A retail company wants to analyze thousands of customer product reviews to determine whether each review is positive, negative, or neutral. Which Azure service and capability should you use?

Show answer
Correct answer: Azure AI Language sentiment analysis
Sentiment analysis in Azure AI Language is designed to classify text by sentiment (positive/negative/neutral), matching the review-scoring requirement. Entity recognition extracts structured items (people, locations, organizations, etc.) but does not label sentiment. Speech to Text transcribes audio to text and does not perform sentiment scoring on written reviews.

2. A support team wants a bot that can answer FAQs using an existing knowledge base of question-and-answer pairs. The team wants minimal custom ML training. Which option best fits?

Show answer
Correct answer: Azure AI Language question answering (QnA pattern) using a knowledge base
Azure AI Language Question Answering is built for FAQ-style scenarios and can use an existing knowledge base with minimal training. Using Azure OpenAI with a base model only is not grounded in the provided QnA source and can hallucinate answers, which is risky for support content. Key phrase extraction identifies important phrases but does not return an answer to a user question.

3. A company needs to translate live chat messages between English and Japanese with low latency. Which Azure AI service should you select?

Show answer
Correct answer: Azure AI Translator
Azure AI Translator is the managed Azure service for language translation, including real-time text translation scenarios such as live chat. Entity recognition identifies entities in text but does not translate. Embeddings are used for semantic search/similarity and retrieval workflows, not for translating text between languages.

4. A healthcare provider wants to generate draft discharge instructions from clinician notes but must reduce unsafe or inappropriate outputs and ensure users know content is AI-generated. Which approach best aligns with the scenario and Responsible AI principles?

Show answer
Correct answer: Use Azure OpenAI with content safety controls and require human review and disclosure of AI-generated content
Azure OpenAI supports generative text and can be paired with safety mitigations (content filtering/safety controls) and governance practices. For medical decision support, Responsible AI principles emphasize reliability & safety, transparency (disclose AI use), and human oversight before use. Key phrase extraction is not generative and does not produce coherent instructions; publishing without review fails safety expectations. Speech to Text only transcribes audio and does not address safe generation or clinical validation, and sending unreviewed content is unsafe.

5. A legal team wants to automatically identify and extract organization names, locations, and dates from contracts to populate a database. Which capability should you use?

Show answer
Correct answer: Azure AI Language named entity recognition (NER)
Named entity recognition in Azure AI Language is designed for extraction of structured entities (for example, organizations, locations, and datetime expressions) from unstructured text, matching the database-population goal. Sentiment analysis classifies opinion polarity and does not extract entities. Azure OpenAI text completion can generate or rewrite text but is not the best-fit managed extraction capability and may produce inconsistent, non-deterministic outputs for structured field extraction compared to NER.

Chapter 6: Full Mock Exam and Final Review

This final chapter is your “dress rehearsal” for AI-900 with a Responsible AI lens. The exam does not reward memorization of feature lists as much as it rewards fast, accurate service selection and correct application of Responsible AI principles (fairness; reliability & safety; privacy & security; inclusiveness; transparency; accountability) in realistic scenarios. Your goal here is to simulate test conditions twice, analyze weak areas using objective-level remediation, and finish with a practical selection matrix and exam-day checklist.

AI-900 questions are often short but loaded with constraints: “real-time,” “offline,” “multilingual,” “no-code,” “PII,” “auditing,” “human review,” “managed service,” “custom model,” “document extraction,” “chatbot,” or “generate text.” Those keywords map directly to Azure services (Azure AI services, Azure Machine Learning, Azure OpenAI, and supporting governance/security tools). In this chapter you’ll also practice a consistent method: identify the workload (vision/NLP/speech/ML/generative), choose the minimum service that meets constraints, then apply the Responsible AI principle that mitigates the stated risk.

Exam Tip: When two answers look plausible, the right one is usually the “closest managed fit” rather than the most powerful platform. For example, choose Azure AI services for common capabilities and Azure Machine Learning when the scenario demands custom training, model registry, or MLOps-style deployment control.

Use the sections below in order: rules and pacing (so you don’t self-sabotage), two mock parts (mixed domains), weak spot analysis, a final selection matrix, and an exam-day checklist. Treat this as the final performance tuning before you schedule or sit the exam.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam rules, pacing plan, and how to review answers

Section 6.1: Mock exam rules, pacing plan, and how to review answers

Your mock exams must feel like the real thing: timed, closed-book, and uninterrupted. AI-900 is not a long exam, but the time pressure comes from context switching across domains (vision → NLP → ML → Responsible AI → generative). Run two mocks on separate days to avoid “learning the answers” rather than learning the reasoning.

Set a pacing plan: first pass answers everything you can in under 45–60 seconds per item; second pass revisits flagged items; final pass is a sanity check for misreads. The most common time sink is rereading a scenario because you didn’t underline constraints the first time.

  • Pass 1: Identify workload + constraints; select the best-fit service; flag uncertainty quickly.
  • Pass 2: For each flagged item, eliminate distractors by matching required capability (e.g., OCR vs. document extraction vs. image classification).
  • Pass 3: Scan for “always/never,” “must,” and compliance cues (PII, encryption, human review, auditability).

Exam Tip: Review mistakes by category, not by question. If you miss three items involving “custom vs prebuilt,” the fix is a rule: “Prebuilt Azure AI services unless the scenario explicitly requires training, bespoke labels, or model lifecycle control.”

When reviewing answers, write a one-line “why” statement that references an exam objective: service selection (vision/NLP/speech), ML fundamentals (training/deployment), generative AI (prompts/capabilities), and Responsible AI principles. If your “why” cannot cite a constraint from the scenario, you likely guessed.

Section 6.2: Mock Exam Part 1 (mixed domains, exam-style)

Section 6.2: Mock Exam Part 1 (mixed domains, exam-style)

Part 1 should mix domains aggressively to mimic the exam’s cognitive switching. Your job is to practice rapid classification: “What kind of problem is this?” (prediction/classification/anomaly detection; OCR/document understanding; translation/sentiment; speech-to-text; prompt-driven generation). Then choose the service tier: Azure AI services for packaged capabilities, Azure Machine Learning for custom training and deployment pipelines, Azure OpenAI for generative tasks.

Anchor your reasoning to keywords. For example, “extract key-value pairs from invoices” points to document processing rather than generic OCR alone; “near real-time transcription” indicates speech services; “detect objects in images” is not the same as “read text in images.” In ML items, watch for the exam’s frequent trap: confusing training (fit a model) with inference (use the trained model).

  • Vision traps: OCR vs. image classification vs. object detection vs. document extraction. If the scenario mentions tables, fields, or forms, think document intelligence/document processing rather than raw OCR.
  • NLP traps: Language detection, entity recognition, sentiment, summarization, and translation are distinct capabilities. If the scenario requires generating new content (not extracting), that is generative AI, not classical NLP analytics.
  • ML traps: Supervised vs. unsupervised; classification vs. regression; evaluation metrics. If the output is a category, it’s classification; if it’s a number, it’s regression.
  • Responsible AI traps: Fairness is not “accuracy.” Privacy & security is not “transparency.” Match the risk: bias → fairness; outages/unsafe outputs → reliability & safety; PII → privacy & security; diverse users → inclusiveness; explanations → transparency; governance → accountability.

Exam Tip: When the scenario mentions “no code/low code,” it’s often nudging you toward a managed Azure AI service or a studio experience rather than building everything in Azure Machine Learning from scratch.

After Part 1, do not immediately read explanations. First, re-answer flagged items using only your constraint notes. This strengthens exam-time discipline: you learn to resolve ambiguity by elimination rather than by “looking it up.”

Section 6.3: Mock Exam Part 2 (mixed domains, exam-style)

Section 6.3: Mock Exam Part 2 (mixed domains, exam-style)

Part 2 should lean heavier into generative AI and Responsible AI integration, because modern AI-900 scenarios increasingly test safe, compliant use of models—not just “which service.” Practice identifying when the scenario is asking for prompt-based generation (Azure OpenAI) versus extraction/analysis (Azure AI Language). Also practice selecting the control that best addresses the risk.

Generative AI questions often hide the true requirement in a single phrase: “draft,” “compose,” “rewrite,” “create variants,” “chat,” “summarize in a new tone,” or “generate code.” Those imply content generation. In contrast, “detect PII,” “classify intent,” “extract entities,” and “sentiment” imply analysis.

Responsible AI concepts are tested as operational decisions. For example: if a system could produce harmful content, reliability & safety is addressed through content filters, grounding to trusted data, and human-in-the-loop review. If the system uses customer data, privacy & security concerns drive data minimization, access control, encryption, and avoiding leaking sensitive data in prompts or logs.

  • Transparency: Users should know they are interacting with AI; provide explanations and limitations; document model behavior and data sources.
  • Accountability: Assign owners, escalation paths, audit logs, and approval workflows. The exam may imply governance when it mentions “audit,” “policy,” or “responsible deployment.”
  • Inclusiveness: Consider accessibility (speech interfaces, captions), multilingual support, and representative testing across user groups.

Exam Tip: If the scenario asks for “the best way to reduce hallucinations,” the correct direction is usually grounding (retrieval with trusted sources), tighter instructions, and validation—not merely “use a bigger model.” Bigger is not a safety control.

Finish Part 2 by writing three remediation rules based on your misses, such as: “If it’s document fields, not just text, pick document intelligence/document extraction,” or “If it’s generating text, it’s Azure OpenAI; if it’s extracting insights from text, it’s Azure AI Language.” These rules become your final-day flash review.

Section 6.4: Score report interpretation and objective-level remediation

Section 6.4: Score report interpretation and objective-level remediation

Your score matters less than your pattern of errors. Interpret results by mapping every incorrect (and guessed-correct) item to one exam objective: AI workloads, ML fundamentals/Azure Machine Learning, computer vision, NLP, generative AI/Azure OpenAI, and Responsible AI principles across scenarios. The goal is to turn “I got it wrong” into “I violated a decision rule.”

Use a simple remediation grid:

  • Concept gap: You didn’t know what a term meant (e.g., regression vs. classification). Fix by reviewing definitions and doing 5 quick classification drills.
  • Service confusion: You knew the workload but chose the wrong service (e.g., OCR vs. document processing). Fix by building a one-page service selection map.
  • Constraint miss: You overlooked a word like “real-time,” “multilingual,” or “PII.” Fix by practicing “constraint highlighting” on each scenario before viewing answers.
  • Responsible AI mismatch: You selected the wrong principle for the risk. Fix by building a risk→principle table and drilling with short scenarios.

Exam Tip: Treat “guessed correctly” as wrong for study purposes. The exam’s distractors are designed so that uncertain knowledge collapses under time pressure.

Remediate in 30–45 minute blocks: one objective per block, one artifact per block (a table, a set of rules, or a mini checklist). End each block by re-solving 5 items of that type without notes. If you cannot articulate why the correct answer is correct in one sentence, you’re not ready in that objective.

Section 6.5: Final review: service selection matrix by domain

Section 6.5: Final review: service selection matrix by domain

This matrix is your final consolidation tool. On AI-900, many wrong answers are “nearly right” services. You win by choosing the smallest service that satisfies the scenario constraints while aligning to Responsible AI expectations.

  • AI Workloads (general): Identify whether the scenario is prediction, perception (vision/speech), language understanding, or generation. Trap: calling everything “ML.” Many scenarios are solved by prebuilt AI services without custom training.
  • Machine Learning on Azure: Use Azure Machine Learning when you need custom model training, experiment tracking, model registry, or managed deployment endpoints. Know basics: training vs. inference; features vs. labels; classification vs. regression. Trap: confusing evaluation with deployment; or selecting ML when a managed AI service already fits.
  • Computer Vision: Use vision services for image classification, object detection, and OCR; use document-focused capabilities when the task is extracting structured fields from forms/invoices. Trap: OCR reads text; it doesn’t inherently understand document structure.
  • NLP: Use language services for sentiment, key phrases, entity recognition, language detection, translation, and conversational analysis patterns. Use speech services for transcription and text-to-speech. Trap: choosing translation when the task is summarization; choosing speech-to-text when the input is already text.
  • Generative AI: Use Azure OpenAI for chat, content drafting, summarization with style change, and code generation. Pair with grounding to trusted data when accuracy matters. Trap: treating generative models like deterministic databases.
  • Responsible AI overlay: Fairness (bias), reliability & safety (robustness and harm reduction), privacy & security (PII and access control), inclusiveness (accessible design), transparency (disclosures/explanations), accountability (governance and audit). Trap: picking “transparency” when the real issue is privacy, or “fairness” when the issue is unsafe output.

Exam Tip: If the question asks “best next step,” the correct answer is often a control or process (human review, monitoring, documentation, access control) rather than a new model or bigger service.

In your final review, rewrite the matrix in your own words on a blank page. Anything you can’t reproduce from memory is a likely exam-day stumble point.

Section 6.6: Exam-day checklist: environment, time management, and mindset

Section 6.6: Exam-day checklist: environment, time management, and mindset

Exam-day performance is mostly execution: reading precisely, managing time, and avoiding avoidable mistakes. Use the checklist below to remove friction so your attention stays on scenario constraints and service selection.

  • Environment: Confirm identity requirements, testing space rules, and stable network (if online). Close apps and notifications. Have acceptable identification ready.
  • Warm-up: Spend 5 minutes reviewing your service selection matrix and the Responsible AI risk→principle table. Do not attempt a full practice set right before the exam.
  • Time management: First pass fast; flag and move. Don’t “argue” with a question—eliminate, choose, and continue. Keep the last minutes for reviewing flags, not rereading everything.
  • Reading discipline: Underline constraints mentally: input type (image/text/audio), real-time vs batch, custom vs prebuilt, compliance/PII, multilingual, and required output (class, number, extracted fields, generated text).

Exam Tip: If you’re torn between two answers, ask: “Which one directly satisfies the stated requirement with the least extra assumptions?” AI-900 favors direct alignment to the scenario, not “future-proofing.”

Mindset: stay literal. Many candidates miss points by inferring requirements that aren’t there (e.g., assuming custom training is needed). When Responsible AI appears, match the principle to the specific risk described. Finally, trust your process: workload → constraints → service → Responsible AI control. That sequence is your exam-day autopilot.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
Chapter quiz

1. A company wants to add a real-time customer support chatbot to its website. The bot must summarize a user’s last three messages and draft a reply, and the company requires a human agent to review and approve the response before it is sent. Which Responsible AI principle is MOST directly addressed by requiring human review?

Show answer
Correct answer: Accountability
Human-in-the-loop review establishes clear oversight and responsibility for outcomes, aligning most directly to accountability. Inclusiveness focuses on supporting diverse users (e.g., accessibility, language support), which is not the main control described. Reliability & safety concerns consistent performance and preventing harmful outputs, but the specific control in the scenario is governance/ownership of decisions rather than a technical safety mechanism.

2. You are given a short exam scenario: “Extract key-value pairs from scanned invoices and return the results as JSON. Use a managed service; do not train a custom model.” Which Azure service best fits the requirements?

Show answer
Correct answer: Azure AI Document Intelligence (Form Recognizer)
Azure AI Document Intelligence is designed for document extraction (invoices, receipts) and can output structured data without requiring custom ML training. Azure Machine Learning is a platform for building/training/deploying custom models and is overkill when a managed extractor is requested. Azure AI Vision image analysis can describe images or detect objects/labels, but it is not the primary managed service for extracting structured fields from documents like invoices.

3. A healthcare organization plans to use a text analytics solution to detect patient sentiment from feedback forms. The data contains personally identifiable information (PII). They want to minimize the risk of exposing sensitive data while still enabling analysis. Which Responsible AI principle is MOST relevant?

Show answer
Correct answer: Privacy & security
Because the scenario centers on protecting PII and reducing exposure of sensitive data, the primary principle is privacy & security. Transparency focuses on making system behavior understandable (e.g., explanations, disclosures), not primarily on data protection. Fairness concerns bias and equitable outcomes across groups; it may matter in healthcare generally, but it does not address the stated risk of PII exposure.

4. An exam question states: “You need to classify images into a set of company-specific defect categories. The categories are unique to your manufacturing process, and you must train with your labeled images and manage deployments.” Which option best matches the ‘closest fit’ service choice?

Show answer
Correct answer: Azure Machine Learning
The need for company-specific labels, custom training, and controlled deployment aligns with Azure Machine Learning (custom model training, registration, and MLOps-style deployment). Azure AI Vision prebuilt analysis is a managed capability for common visual features and is not intended for training bespoke defect classifiers in this framing. Azure AI Translator is for language translation and is unrelated to image classification.

5. A team deploys an AI solution that helps screen job applicants. After launch, they discover the model’s recommendations differ significantly across demographic groups. Which action best aligns with Responsible AI practice for addressing this issue?

Show answer
Correct answer: Evaluate and mitigate bias using fairness metrics and adjust the data/model, then re-test before redeploying
Disparate outcomes across demographic groups point to a fairness risk; the appropriate response is to measure bias with fairness metrics, investigate root causes (data imbalance, feature leakage, label bias), mitigate, and re-validate prior to redeployment. Increasing model complexity may improve overall accuracy but can worsen or hide bias and does not address the fairness constraint. Removing logging reduces auditability and accountability; it does not fix bias and can hinder required monitoring and governance.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.