HELP

+40 722 606 166

messenger@eduailast.com

Microsoft AI Fundamentals for Non-Technical Pros (AI-900)

AI Certification Exam Prep — Beginner

Microsoft AI Fundamentals for Non-Technical Pros (AI-900)

Microsoft AI Fundamentals for Non-Technical Pros (AI-900)

Understand Azure AI concepts and pass AI-900 with focused practice.

Beginner ai-900 · microsoft · azure · ai-fundamentals

Course goal: pass Microsoft AI-900 with confidence

This Edu AI exam-prep course is designed for non-technical professionals preparing for the Microsoft AI-900: Azure AI Fundamentals certification exam. You’ll learn the core concepts Microsoft expects you to recognize on exam day—without requiring a software engineering background—then you’ll validate your understanding with exam-style practice that mirrors how AI-900 questions are phrased.

AI-900 focuses on recognizing AI concepts, choosing the right Azure AI capabilities for common business scenarios, and understanding responsible AI fundamentals. This course is built as a 6-chapter “book,” so you can progress in a predictable, beginner-friendly sequence and quickly identify weak areas before scheduling your exam.

Aligned to the official AI-900 exam domains

The blueprint maps directly to the official objectives by name, covering:

  • Describe AI workloads
  • Fundamental principles of machine learning on Azure
  • Computer vision workloads on Azure
  • NLP workloads on Azure
  • Generative AI workloads on Azure

Across Chapters 2–5, you’ll learn how to interpret real-world prompts such as “Which workload is this?”, “Which service should you use?”, and “Which metric best evaluates the model?”—the exact kind of decision-making the AI-900 exam checks.

How the 6 chapters are structured

Chapter 1 sets you up for success: how to register and schedule, what to expect from the testing experience, how scoring works, and how to study efficiently as a beginner. You’ll also get a practical plan for using short daily sessions to retain terminology and service capabilities.

Chapters 2–5 each focus on one (or two closely related) exam domains. The emphasis is on clear definitions, scenario recognition, and service selection—so you can answer questions even when Microsoft changes the story details. Each chapter ends with exam-style practice milestones to reinforce the objective names and the “best answer” logic.

Chapter 6 provides a full mock exam split into two parts, followed by weak-spot analysis and a final review sprint. You’ll leave with a checklist you can use the day before and the day of the exam.

Why this course helps you pass

  • Beginner-first explanations: No prior certification experience required; concepts are taught from the ground up.
  • Objective-based coverage: Lessons and practice are organized to match Microsoft’s AI-900 domain language.
  • Exam-style practice: Scenario-based questions that train you to choose the best option, not just recall terms.
  • Practical review workflow: A mock exam and a repeatable method to close gaps quickly.

Get started

If you’re ready to begin, you can Register free and start studying today. Prefer to compare options first? You can also browse all courses on Edu AI.

What You Will Learn

  • Describe AI workloads (including responsible AI) and common Azure AI scenarios
  • Explain fundamental principles of machine learning on Azure, including training, evaluation, and AutoML
  • Identify computer vision workloads on Azure and choose appropriate Azure AI Vision services
  • Identify NLP workloads on Azure and choose appropriate Azure AI Language and Speech capabilities
  • Describe generative AI workloads on Azure, including Azure OpenAI concepts and responsible use

Requirements

  • Basic IT literacy (web apps, cloud basics, and common business software)
  • No prior Microsoft certification experience required
  • A computer with reliable internet access for practice quizzes and review

Chapter 1: AI-900 Exam Orientation and Study Strategy

  • Understand AI-900 purpose and audience fit
  • Register, schedule, and take the exam (online or test center)
  • How scoring, question types, and time management work
  • Build your 2-week and 4-week study plans
  • Set up practice routine and exam-day readiness

Chapter 2: Describe AI Workloads (and Responsible AI)

  • Recognize common AI workload types and business use cases
  • Differentiate AI, ML, deep learning, and generative AI at a high level
  • Apply responsible AI concepts to real scenarios
  • Domain practice set: AI workloads and responsible AI

Chapter 3: Fundamental Principles of Machine Learning on Azure

  • Understand core ML concepts: data, features, labels, and model lifecycle
  • Distinguish supervised vs unsupervised learning and common algorithms
  • Interpret evaluation metrics and avoid common pitfalls
  • Domain practice set: ML principles on Azure

Chapter 4: Computer Vision Workloads on Azure

  • Identify core vision tasks and the right Azure service fit
  • Understand image analysis, OCR, and video/face scenarios at a high level
  • Know typical inputs/outputs, constraints, and responsible considerations
  • Domain practice set: computer vision workloads

Chapter 5: NLP and Generative AI Workloads on Azure

  • Choose the right NLP task: sentiment, key phrases, entities, summarization
  • Understand speech scenarios: speech-to-text, text-to-speech, translation
  • Explain generative AI basics and Azure OpenAI use cases safely
  • Domain practice set: NLP + generative AI workloads

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Final Review Sprint
  • Exam Day Checklist

Jordan Whitaker

Microsoft Certified Trainer (MCT)

Jordan Whitaker is a Microsoft Certified Trainer who has helped hundreds of learners prepare for Microsoft Fundamentals exams. Jordan specializes in translating Azure AI concepts into practical, non-technical explanations aligned to official exam objectives.

Chapter 1: AI-900 Exam Orientation and Study Strategy

AI-900 (Microsoft Azure AI Fundamentals) is designed to confirm that you can speak confidently about AI workloads and choose appropriate Azure AI services—without needing to be a data scientist or a developer. In other words, the exam rewards clear conceptual thinking and good service-selection judgment more than hands-on coding. This chapter orients you to the exam’s purpose, logistics, and the study behaviors that reliably produce a passing score.

As you prepare, keep the course outcomes in view: you must describe common AI workloads (including responsible AI), explain core machine learning ideas (training, evaluation, and AutoML), identify computer vision and NLP workloads and map them to Azure AI services, and describe generative AI workloads and Azure OpenAI concepts. The exam is “fundamentals,” but it is not vague—expect precise wording, scenario cues, and distractors that look plausible if you haven’t learned the official Azure terminology.

You’ll also build a plan (2-week sprint or 4-week steady pace), set up a practice routine, and learn exam-day readiness habits. The goal is to prevent a common failure mode: studying “interesting AI topics” broadly, instead of studying the AI-900 objectives with deliberate recall and service mapping.

Practice note for Understand AI-900 purpose and audience fit: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register, schedule, and take the exam (online or test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How scoring, question types, and time management work: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build your 2-week and 4-week study plans: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Set up practice routine and exam-day readiness: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand AI-900 purpose and audience fit: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register, schedule, and take the exam (online or test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How scoring, question types, and time management work: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build your 2-week and 4-week study plans: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Set up practice routine and exam-day readiness: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: What AI-900 validates and how it maps to Azure AI Fundamentals

Section 1.1: What AI-900 validates and how it maps to Azure AI Fundamentals

AI-900 validates foundational literacy: recognizing AI workload types, understanding basic machine learning terms, and selecting the right Azure AI capability for a business scenario. For non-technical professionals, the exam is especially aligned to real workplace decisions: “Which service should we use?” and “What risks or responsible AI considerations should we address?”

Mentally map the exam into four recurring content buckets that reappear across objectives and questions: (1) AI workloads and responsible AI, (2) machine learning fundamentals (training vs. inference, evaluation metrics, AutoML), (3) computer vision and NLP workloads with Azure AI services, and (4) generative AI concepts and responsible use (including Azure OpenAI). The exam typically tests that you can differentiate workloads by their inputs/outputs (text, images, audio, tabular data) and by the task (classification, regression, clustering, object detection, OCR, sentiment analysis, speech-to-text, etc.).

Exam Tip: When a prompt describes a scenario, underline the “data type” (text/image/audio/tabular) and the “desired outcome” (predict a number, categorize, extract entities, summarize, generate). Those two clues often eliminate 2–3 distractors immediately.

Common traps include confusing machine learning with rules-based automation, mixing up training (model learns from labeled or unlabeled data) with inference (model applies what it learned), and selecting a service family incorrectly (for example, choosing a general ML platform when the question is clearly about a prebuilt Azure AI service such as Vision, Language, or Speech). Also expect responsible AI themes: fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. You won’t be asked to implement governance, but you can be asked to identify what principle is at risk in a scenario.

Audience fit: AI-900 is appropriate if you’re in business analysis, project management, sales, marketing, operations, or leadership roles that interact with AI projects. Technical depth is limited; however, vocabulary precision is required. You should be able to explain concepts using Microsoft’s terms, because the exam’s “correct” option is often the one that matches how Microsoft positions the service.

Section 1.2: Registration steps, pricing considerations, and scheduling with Pearson VUE

Section 1.2: Registration steps, pricing considerations, and scheduling with Pearson VUE

Scheduling is part of your study strategy because a firm date improves follow-through. AI-900 is delivered through Pearson VUE, available online (proctored) or at a test center. Your registration flow is typically: sign in with your Microsoft account, select the exam, choose your delivery method, and book an appointment. Plan your date so you finish content review at least 3–5 days before the exam, leaving time for targeted practice and weak-area repair.

Pricing can vary by region, and discounts may apply through programs such as Microsoft student discounts, employer vouchers, or event-based vouchers. Treat the payment step as a commitment device: if you’re prone to endless “one more week of studying,” booking and paying can prevent drift. If your employer sponsors certification, confirm voucher rules and rescheduling policies early.

Online proctoring demands a clean environment, valid ID, stable internet, and a compatible device. A common trap is ignoring the system test: camera permissions, network restrictions, and corporate VPN policies can derail your exam. Test center delivery reduces tech risk and can be better if your home environment is unpredictable.

Exam Tip: Choose your delivery mode based on controllability, not convenience. If you cannot guarantee a quiet room for 90+ minutes (including check-in), book a test center. If you go online, run Pearson VUE’s system test a few days before and again on exam day.

Rescheduling and cancellation rules matter. Build in a buffer: book for the end of week 2 (for the 2-week plan) or end of week 4 (for the 4-week plan), but ensure you have at least one alternative slot available within your target window in case illness or work travel hits.

Section 1.3: Exam format: question styles, case-style items, and exam UI expectations

Section 1.3: Exam format: question styles, case-style items, and exam UI expectations

AI-900 uses a mix of question formats intended to test recognition and decision-making rather than long calculations. Expect classic multiple-choice and multi-select items, plus scenario-based prompts where you pick the best service or concept. Some items are “case-style,” meaning you read a short business context and answer several questions from it. The UI will show a question navigator, remaining time, and review capabilities (not all questions allow review depending on the item type).

Time management is straightforward if you avoid overthinking. Fundamentals exams typically reward first-pass accuracy: answer what the question asks, not what you wish it asked. If the scenario is about extracting text from images, don’t drift into a discussion of model training; if it’s about predicting a number from historical data, that is a regression cue, not classification.

Exam Tip: Watch for qualifiers like “best,” “most cost-effective,” “prebuilt,” or “custom.” “Prebuilt” usually points to Azure AI services (Vision/Language/Speech) rather than building and training your own model. “Custom” and “train” often signal Azure Machine Learning (or custom features inside the service family, depending on the objective wording).

Common exam traps in UI and wording include: (1) selecting more options than required in multi-select questions, (2) missing that the scenario requires a responsible AI action (e.g., transparency or privacy), and (3) choosing a general-purpose tool when a specialized service exists. Practice reading the last line first—often it states the exact task: classify, detect, recognize, extract, summarize, generate, translate, or transcribe.

Case-style items can induce fatigue because you re-scan the same text. Use a simple method: on first read, capture three notes—data type, goal, and constraints (latency, cost, compliance, interpretability). Then answer each sub-question using those notes rather than rereading the entire case.

Section 1.4: Scoring, passing expectations, and how to interpret results

Section 1.4: Scoring, passing expectations, and how to interpret results

Microsoft exams generally use a scaled score model, with 700 as the typical passing threshold on a 1000-point scale. Scaled scoring means not every question is worth the same amount, and the score reflects exam form difficulty adjustments. The practical takeaway is that chasing perfection is inefficient; you want reliable competence across all objective areas.

After the exam, you receive a score report showing performance by objective domain. Use it diagnostically: if you pass, it identifies weaker domains for real-world learning; if you fail, it tells you where to focus for a retake. Many candidates misinterpret domain bars as precise percentages; treat them as directional guidance rather than exact measurements.

Exam Tip: Aim for “no weak domains.” A strong performance in one domain rarely compensates for a very weak performance in another because questions are distributed to cover the objective set. Your study plan should always include at least light coverage of every domain each week.

Expectations: AI-900 does not require deep math, but it does expect you to understand evaluation at a high level (accuracy vs. precision/recall concepts, and why evaluation matters), training vs. inference, and responsible AI principles. Another frequent scoring trap is conflating generative AI concepts with classical ML: generative AI questions may focus on safe deployment, prompt-based interaction, and appropriate use cases, not on “training a neural network from scratch.”

Interpreting results also includes knowing what not to do: don’t rebuild your study plan solely around memorizing answers from a single practice set. The exam measures objective understanding and service selection; memorization breaks when Microsoft refreshes items or rephrases scenarios.

Section 1.5: Study strategy for beginners: spaced repetition, recall, and domain weighting

Section 1.5: Study strategy for beginners: spaced repetition, recall, and domain weighting

Your goal is to convert unfamiliar terms into quick recognition under time pressure. The most efficient approach is a combination of spaced repetition (reviewing key ideas over multiple days) and active recall (forcing yourself to retrieve the concept without looking). Passive reading alone feels productive but often fails on scenario questions.

Two proven pacing options:

  • 2-week plan (accelerated): Days 1–3 cover AI workloads + responsible AI; Days 4–6 cover ML fundamentals and AutoML; Days 7–9 cover Vision + Language + Speech mapping; Days 10–11 cover generative AI and Azure OpenAI concepts; Days 12–14 are for mixed practice, weak-area repair, and final review.
  • 4-week plan (steady): Each week targets one major domain while revisiting prior domains 2–3 times using short recall drills. Week 4 is practice-heavy with objective-level remediation.

Domain weighting matters, but don’t over-optimize early. Beginners commonly overweight what feels intuitive (e.g., “chatbots are fun”) and underweight fundamentals like evaluation, responsible AI, and service boundaries. Instead, build a “service selection map” you can recite: when to use Azure AI Vision vs. Azure AI Language vs. Azure AI Speech vs. Azure Machine Learning vs. Azure OpenAI. Then attach 2–3 canonical use cases to each.

Exam Tip: Create a one-page “workload-to-service” table and review it daily during the final week. If you can map 15–20 scenario cues to services quickly, your score rises even if you forget some definitions.

Spaced repetition routine: after each study session, write 8–12 flash prompts in your own words (e.g., “When would I choose prebuilt OCR vs. custom model training?”). Review them 1 day later, 3 days later, and 7 days later. Active recall routine: close the book and explain aloud the difference between classification and regression, or between OCR and object detection, as if briefing a stakeholder.

Section 1.6: Using Microsoft Learn + practice tests: how to study by objective names

Section 1.6: Using Microsoft Learn + practice tests: how to study by objective names

Microsoft Learn is your primary source because it mirrors the exam objective language. The highest-yield study method is to study “by objective name,” not by random browsing. Start with the official skills outline, then align each Learn module you complete to a specific objective such as describing AI workloads and responsible AI, explaining ML training/evaluation/AutoML, identifying computer vision scenarios and Azure AI Vision capabilities, identifying NLP scenarios and Azure AI Language/Speech capabilities, and describing generative AI workloads with Azure OpenAI concepts.

Practice tests are most valuable when used as a diagnostic tool. Take an initial baseline practice test early (even if you score poorly) to reveal blind spots in terminology and service mapping. Then switch to targeted practice: for every missed item, write down (1) the objective it relates to, (2) the key cue in the scenario, and (3) the reason the correct answer is correct in Microsoft terms.

Exam Tip: Don’t just review why your answer was wrong—also write why each distractor is wrong. AI-900 distractors often represent a nearby service or concept (e.g., a general ML platform vs. a prebuilt AI service). Learning the boundary lines is the exam skill.

Build your practice routine around short cycles: Learn module → 10–15 minutes of recall notes → a small set of objective-aligned practice items → remediation. Avoid marathon practice sessions that encourage guessing. The exam rewards calm, consistent recognition. For exam-day readiness, simulate the full experience at least once: quiet environment, timed run, no notes, and a strict review pass. If you’re taking the exam online, also rehearse your physical setup (camera angle, desk clearance, ID ready) so logistics do not consume mental bandwidth.

Finally, remember the meta-skill AI-900 tests: choosing the right capability responsibly. When in doubt during practice, return to the objective names. If you can explain the objective in plain language and map it to Azure services, you are studying the way the exam is written.

Chapter milestones
  • Understand AI-900 purpose and audience fit
  • Register, schedule, and take the exam (online or test center)
  • How scoring, question types, and time management work
  • Build your 2-week and 4-week study plans
  • Set up practice routine and exam-day readiness
Chapter quiz

1. You are advising a marketing manager who wants an entry-level Microsoft certification to validate that they can discuss AI workloads and select appropriate Azure AI services, without writing code. Which certification best matches this goal?

Show answer
Correct answer: AI-900: Microsoft Azure AI Fundamentals
AI-900 is a fundamentals exam focused on conceptual understanding of AI workloads and mapping them to Azure AI services—intended for non-technical or broadly technical audiences. DP-100 targets data scientists and requires hands-on ML implementation skills. AZ-204 is a developer-focused exam centered on building solutions in Azure and is not aligned to a non-coding AI fundamentals outcome.

2. A candidate says, "I’m studying AI-900 by reading general articles about AI because it’s fundamentals, so exact Azure service names aren’t important." What is the biggest risk with this approach based on how AI-900 questions are written?

Show answer
Correct answer: They may miss precise Azure terminology and service-selection cues used in scenario questions and distractors
AI-900 rewards clear conceptual thinking and correct service selection, and questions often include precise wording and plausible distractors that require knowing official Azure AI terminology. It does not require coding implementation (so B is incorrect). It is not a math-heavy calculus/statistics exam (so C is incorrect), though it does include basic ML concepts like training and evaluation.

3. You have 14 days until your AI-900 exam. You can study about 60–90 minutes per day. Which plan best aligns with the chapter’s recommended study strategy?

Show answer
Correct answer: Use a 2-week sprint plan focused on the AI-900 objectives, deliberate recall, and practice questions mapped to services
A 2-week sprint is appropriate for a short timeline and should stay tightly aligned to AI-900 objectives, using deliberate recall and practice that reinforces workload-to-service mapping. Broad exploration (B) risks missing exam objectives and the specific Azure terminology the exam expects. Avoiding practice (C) undermines exam readiness because AI-900 uses scenario cues and distractors that are best handled through active recall and targeted practice.

4. During the exam, you notice some questions are long and scenario-based with multiple plausible answers. What time-management behavior best matches the chapter’s guidance on scoring and question types?

Show answer
Correct answer: Answer what you can efficiently, flag harder items, and manage time across the full set rather than getting stuck on one question
Fundamentals exams often include scenario-based items and distractors; effective time management means progressing steadily, using flags/review strategically, and not fixating on a single item. Certification exams do not typically weight early questions more heavily (so B is incorrect). Scenario-based questions do count toward scoring (so C is incorrect) and are common in AI-900 to test service-selection judgment.

5. A company wants its staff to be "exam-day ready" for AI-900. Which preparation activity most directly supports readiness (beyond learning content) as emphasized in the chapter?

Show answer
Correct answer: Run timed practice sets and rehearse the exam-day logistics (test center vs. online), including check-in requirements
Exam-day readiness includes practical logistics (online vs. test center), check-in expectations, and timed practice to build pacing and reduce surprises. AI-900 does not require building coded pipelines end-to-end (B), making that preparation misaligned with the exam’s fundamentals focus. Memorizing definitions without practice (C) fails to develop the service-mapping and scenario interpretation skills that AI-900 commonly tests.

Chapter 2: Describe AI Workloads (and Responsible AI)

This chapter maps directly to the AI-900 objective area that asks you to describe AI workloads and recognize common Azure AI scenarios—while consistently applying Responsible AI. As a non-technical professional, your exam success depends less on coding details and more on choosing the right workload type (prediction vs classification vs detection vs generation), knowing when AI is appropriate, and recognizing which Azure AI service family fits a scenario.

AI-900 questions are often written as short business scenarios with a “best answer” choice. The exam expects you to identify the workload first, then select the right tool family, and finally show awareness of responsible use. A common trap is to overfocus on product names before you understand the workload. Another is confusing “classification” (a label) with “detection” (finding where something is) or “prediction” (a numeric value or future outcome). You’ll practice a reliable approach in Section 2.6.

Throughout this chapter, keep a simple mental checklist: (1) What is the input type (text, image, audio, tabular data)? (2) What is the output type (label, number, location, generated content)? (3) Is the task deterministic rules or pattern-based learning? (4) What risks exist (fairness, privacy, safety), and how would you mitigate them? That’s the exam mindset.

Practice note for Recognize common AI workload types and business use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate AI, ML, deep learning, and generative AI at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI concepts to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: AI workloads and responsible AI: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize common AI workload types and business use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate AI, ML, deep learning, and generative AI at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI concepts to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: AI workloads and responsible AI: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize common AI workload types and business use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: Official objective—Describe AI workloads: prediction, classification, detection, generation

Section 2.1: Official objective—Describe AI workloads: prediction, classification, detection, generation

The AI-900 exam repeatedly tests whether you can name the correct workload based on the output the system produces. Four workload verbs show up constantly: prediction, classification, detection, and generation. Learn to spot them in a sentence.

Prediction means estimating a numeric value or a future outcome from patterns in historical data. Think “forecast,” “estimate,” or “risk score.” Examples: predicting next month’s demand, estimating house price, forecasting call volume. On the exam, prediction often aligns with regression-style thinking (even if the term “regression” isn’t used).

Classification assigns a category label. If the output is one of several named groups, it’s classification: “spam vs not spam,” “fraud vs legitimate,” “positive/neutral/negative sentiment,” “document type A/B/C.” The trap: learners confuse “classification” with “prediction” because both are “predicting.” In exam wording, if the output is a discrete label, call it classification.

Detection identifies the presence and often the location of something. In vision scenarios, detection typically means drawing bounding boxes around objects (object detection) or finding anomalies/defects. In security or IoT, “anomaly detection” can also be framed as detection. The exam trap is mixing “image classification” (what is it?) with “object detection” (where is it?). If the question mentions counting, locating, bounding boxes, or identifying items in a scene, think detection.

Generation creates new content: text, images, code, summaries, or responses. This is the generative AI family. If the requirement is “draft,” “write,” “summarize,” “translate with style,” “compose,” or “create,” it’s generation. Exam Tip: If the output is novel content rather than a label/number, choose “generation” even if the prompt includes analysis (e.g., “summarize customer feedback” is generation, not classification).

Most scenario questions become easy once you name the workload type first; the right Azure service family usually follows naturally.

Section 2.2: When to use AI vs traditional software rules (decision criteria and examples)

Section 2.2: When to use AI vs traditional software rules (decision criteria and examples)

AI-900 expects you to know when AI is the right tool and when simple rules or conventional software is better. The key divider is whether the problem can be expressed as explicit, stable logic or whether it requires learning patterns from data.

Use traditional rules when: the logic is clear, requirements change infrequently, and you can enumerate conditions. Examples: calculating sales tax, validating a form field, routing tickets by exact keywords, enforcing a password policy. These are deterministic and auditable.

Use AI/ML when: the logic is hard to write down, inputs are noisy (images, speech, natural language), or patterns shift and need adaptation. Examples: identifying damaged products in photos, detecting fraudulent transactions based on many signals, extracting meaning from customer emails, or predicting churn from customer behavior.

Decision criteria the exam likes: (1) Data availability—do you have enough representative data? (2) Tolerance for error—AI is probabilistic; are you okay with a confidence score and occasional mistakes? (3) Explainability needs—do you need a clear rule trail for compliance? (4) Cost/benefit—is building/training worth it versus simpler automation?

Exam Tip: Watch for questions that “smuggle” a rules problem into AI wording. If the scenario says the outcome is based on a fixed policy (“If customer is in region X and order > $Y, approve”), the best answer is usually not ML. A common trap is choosing AI because it sounds modern, even when rules are sufficient.

Also recognize hybrid designs: rules can pre-filter cases, and AI can handle ambiguous or unstructured inputs. The exam may reward answers that correctly place AI where uncertainty exists and keep deterministic steps as rules.

Section 2.3: Azure AI service families overview: Vision, Language, Speech, Decision, Search, OpenAI

Section 2.3: Azure AI service families overview: Vision, Language, Speech, Decision, Search, OpenAI

AI-900 is not a memorization test of every product SKU, but it does expect you to map a workload to the correct Azure AI service family. Think in families first, then pick the best fit.

Vision covers image and video understanding: image analysis, OCR (reading text), object detection, and basic face-related analysis (where permitted). If the input is a photo, scanned document, or video frame, Vision is the default family. Typical phrasing: “identify items in an image,” “read text from receipts,” “detect defects.”

Language handles text understanding: sentiment analysis, key phrase extraction, entity recognition, summarization, and classification of text. If the input is emails, chat logs, documents, or reviews—and the output is insights about the text—Language is usually right.

Speech covers audio: speech-to-text, text-to-speech, speech translation, and speaker-related capabilities. If microphones, call recordings, or voice interfaces are involved, think Speech.

Decision is for recommendation and personalization-style choices. The exam may describe “show the best product to each user” or “rank options based on user behavior.” In modern Azure, some older branded services changed over time; the test goal is recognizing the workload (recommendation/ranking/decisioning) rather than a legacy name.

Search supports indexing and retrieval over enterprise content, often enhanced with AI enrichment and natural language querying. If the scenario is “find answers in company documents,” “search a knowledge base,” or “retrieve relevant passages,” Search is a strong candidate—often paired with Language or OpenAI for conversational experiences.

OpenAI (Azure OpenAI) targets generative AI: chat, content creation, summarization, code generation, and embeddings for semantic similarity. If the scenario describes generating responses, drafting content, or building a chat assistant, Azure OpenAI is typically the best family.

Exam Tip: Many incorrect options are “near misses” (e.g., using Search when the requirement is sentiment classification, or using Language when the requirement is speech transcription). Anchor on the input modality first (image/text/audio) to avoid this trap.

Section 2.4: Responsible AI fundamentals: fairness, reliability & safety, privacy & security, inclusiveness, transparency, accountability

Section 2.4: Responsible AI fundamentals: fairness, reliability & safety, privacy & security, inclusiveness, transparency, accountability

Responsible AI is not an “extra” topic on AI-900—it’s woven into scenario questions. You’re expected to recognize risks and select practices that reduce harm. Microsoft frames Responsible AI around six principles, and the exam often uses their exact names.

Fairness: AI should treat groups equitably. Exam scenarios often involve hiring, lending, insurance, or school admissions. The trap is assuming “remove protected attributes” automatically fixes bias; correlated variables can still encode unfairness. Look for mitigations like representative data, bias evaluation, and ongoing monitoring.

Reliability & safety: Systems should behave as intended across conditions, including edge cases. For example, a vision model trained on daylight images may fail at night. You may see questions about failover, human review, confidence thresholds, and testing across environments.

Privacy & security: Protect personal data and prevent misuse. Scenarios include medical records, voice recordings, or customer chats. Correct answers often mention access control, data minimization, encryption, and avoiding logging sensitive prompts/output unnecessarily.

Inclusiveness: Ensure the system works for people with different abilities, languages, accents, or devices. For instance, speech systems must handle diverse accents; vision systems should work with different lighting and skin tones. Inclusiveness is frequently confused with fairness; inclusiveness is about usability and accessibility across populations and contexts.

Transparency: Users should understand AI is involved and what it can/can’t do. Good answers include communicating limitations, showing confidence scores, explaining data usage, and labeling AI-generated content where appropriate.

Accountability: Humans remain responsible. The exam likes governance ideas: clear ownership, audit trails, escalation paths, and human-in-the-loop for high-impact decisions.

Exam Tip: When you see a high-stakes domain (health, finance, employment), prioritize answers that add human review, documentation, and monitoring—not just “train a better model.” Responsible AI answers tend to be process- and policy-oriented, not purely technical.

Section 2.5: Building blocks: data, features, labels, prompts, endpoints, and evaluation basics

Section 2.5: Building blocks: data, features, labels, prompts, endpoints, and evaluation basics

Even for non-technical candidates, AI-900 expects basic vocabulary for how AI solutions are built and assessed. Many questions hide the correct choice behind these foundational terms.

Data is the raw input collected from the world: transactions, images, support tickets, sensor readings. Features are the measurable attributes used for learning (e.g., “purchase frequency,” “time on site,” “image pixel patterns,” “text embeddings”). Labels are the correct answers in supervised learning (e.g., “fraud/not fraud,” “product category,” “defect type”). A classic trap: confusing features with labels. If it’s the thing you want to predict, it’s a label; if it helps you predict, it’s a feature.

Prompts are instructions or inputs you give to a generative model. Prompting is not the same as training; it’s guiding a pre-trained model at runtime. On the exam, if the scenario mentions “system message,” “instructions,” or “few-shot examples,” that’s prompt engineering territory.

Endpoints are how apps consume models/services—think of a deployed URL/API that receives inputs and returns outputs. If a question says “integrate into an app,” “call from a workflow,” or “deploy for consumption,” endpoints are implied.

Evaluation basics: You don’t need advanced math, but you must know that models are evaluated on data they haven’t seen (test/validation) and compared using metrics appropriate to the task. For classification, accuracy is common, but imbalanced scenarios push you toward precision/recall thinking (without necessarily requiring formulas). For generative AI, evaluation often includes quality, groundedness, safety, and human review.

Exam Tip: If the scenario demands “measure performance before release” or “compare two approaches,” choose options that mention evaluation on holdout data, monitoring in production, and feedback loops. The exam punishes “train once and forget.”

Section 2.6: Exam-style practice: scenario selection and “best answer” reasoning for AI workloads

Section 2.6: Exam-style practice: scenario selection and “best answer” reasoning for AI workloads

This section teaches the exam skill behind many AI-900 items: selecting the “best answer” when multiple choices sound plausible. Your goal is to follow a repeatable reasoning path rather than guessing.

Step 1: Identify the workload verb from the output. If the output is a category, it’s classification. If it’s a number or forecast, it’s prediction. If it’s “where in the image/audio stream,” it’s detection. If it’s newly written content, it’s generation. This single step eliminates many distractors.

Step 2: Anchor on the input modality. Image/video implies Vision; text implies Language; audio implies Speech; enterprise retrieval implies Search; generative drafting implies OpenAI. If the scenario mixes modalities (e.g., “call center recordings summarized”), you may need Speech first (transcribe) and then Language/OpenAI (summarize). The exam often rewards recognizing the primary service family, even if the full pipeline isn’t listed.

Step 3: Check if rules are enough. If the requirement is fixed logic or simple thresholds, traditional software is often the best answer. If it requires pattern recognition from messy data, AI is appropriate. Distractors frequently present ML for problems that are clearly rule-based.

Step 4: Apply Responsible AI as a tie-breaker. If two answers seem correct, prefer the one that mentions safeguards: human review for high impact, privacy controls for sensitive data, transparency to users, or monitoring for drift and safety. Responsible AI isn’t just ethics—it’s exam points.

Common traps to avoid:

  • Classification vs detection: “Identify the type of object” (classification) vs “locate and count objects” (detection).
  • Search vs OpenAI: If the task is retrieving the right documents, that’s Search; if it’s composing a response, that’s OpenAI (often with Search as grounding).
  • Language vs Speech: If the input is spoken audio, you need Speech capabilities before text analytics.
  • Prompting vs training: If you’re changing behavior by instructions/examples at runtime, that’s prompting; if you’re fitting a model to labeled data, that’s training.

Exam Tip: When you see words like “best,” “most appropriate,” or “first,” the exam is asking for prioritization. Choose the option that directly satisfies the requirement with the least complexity and the clearest alignment to the workload verb and modality.

By applying this method consistently, you’ll turn scenario questions into a structured selection process—exactly what AI-900 is designed to assess for non-technical professionals.

Chapter milestones
  • Recognize common AI workload types and business use cases
  • Differentiate AI, ML, deep learning, and generative AI at a high level
  • Apply responsible AI concepts to real scenarios
  • Domain practice set: AI workloads and responsible AI
Chapter quiz

1. A retail company wants to automatically route incoming customer emails into one of three queues: Billing, Technical Support, or Returns. The input is email text, and the output is a single category label. Which AI workload type best fits this requirement?

Show answer
Correct answer: Text classification
This is a classification workload because the system assigns one of several discrete labels to a text input. Object detection is used to locate objects within images (it outputs locations like bounding boxes), not to label emails. Regression prediction outputs a numeric value (for example, predicting revenue or demand), not a category such as Billing or Returns.

2. A manufacturer wants to analyze images from a quality-control camera to find defective parts and draw a box around each defect in the image. Which AI workload should you identify first?

Show answer
Correct answer: Object detection
Object detection is the correct workload because it identifies what is present and where it is in an image (for example, bounding boxes around defects). Image classification assigns a single label to the entire image (for example, defective vs not defective) and does not provide locations. Clustering is an unsupervised technique to group similar items and is not the typical workload for locating defects in an image.

3. A utility company wants to estimate next month’s electricity usage for each customer based on historical monthly usage and weather data. The output is a numeric value. Which workload type is most appropriate?

Show answer
Correct answer: Regression
Estimating a numeric value (future electricity usage) is a regression workload. Classification produces discrete labels (for example, high/medium/low), which is not what is requested. Generative AI focuses on creating new content such as text, images, or code rather than predicting a numeric value from tabular historical data.

4. A nonprofit plans to use an AI model to help screen job applicants. They are concerned the model may disadvantage certain demographic groups due to biased historical data. Which Responsible AI principle is most directly being addressed by this concern?

Show answer
Correct answer: Fairness
The concern about disadvantaging certain groups maps most directly to the fairness principle (avoiding bias and disparate impact). Reliability and safety focuses on consistent, robust operation and avoiding harmful failures, not primarily on bias across groups. Transparency is about making AI decisions understandable and providing appropriate explanations, which is important but is not the core issue described.

5. A marketing team wants a system that can draft new product descriptions and slogans in the company’s brand voice. They want original text based on a short prompt. Which statement best differentiates the needed approach from traditional machine learning?

Show answer
Correct answer: Generative AI produces new content, while traditional ML commonly predicts labels or numeric values from existing data.
Generative AI is used to create new content (for example, drafting product descriptions) from prompts, whereas traditional ML scenarios in AI-900 often focus on predicting a label (classification) or a number (regression) from input data. Deep learning does not mean 'no training data' or 'fixed rules'; it is a subset of machine learning that typically requires significant training data and learns patterns. Classification outputs a label and is not intended to generate original long-form text.

Chapter 3: Fundamental Principles of Machine Learning on Azure

AI-900 expects you to recognize machine learning (ML) as a specific AI workload focused on learning patterns from data to make predictions or decisions. This chapter connects the “what” (core ML concepts like features, labels, and evaluation) to the “where” (Azure Machine Learning capabilities) and the “how to pass” (how exam questions are worded and the common traps).

Non-technical professionals often think ML is a single step: “feed data in, get answers out.” The exam tests that you understand the lifecycle: data preparation, training, evaluation, deployment, and then inference (using the model). You’re also expected to distinguish supervised learning (labels provided) from unsupervised learning (no labels), pick appropriate evaluation metrics for a scenario, and identify when AutoML or a low-code tool is the best fit.

As you read, watch for how questions hint at the right learning type (classification vs regression vs clustering) and the right metric (accuracy vs F1 vs RMSE). Many incorrect options are “almost right” but mismatched to the business goal or data type.

Practice note for Understand core ML concepts: data, features, labels, and model lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Distinguish supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Interpret evaluation metrics and avoid common pitfalls: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: ML principles on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand core ML concepts: data, features, labels, and model lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Distinguish supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Interpret evaluation metrics and avoid common pitfalls: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: ML principles on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand core ML concepts: data, features, labels, and model lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Distinguish supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: Official objective—Fundamental principles of ML on Azure: training vs inference

The AI-900 objective in this chapter is to explain fundamental ML principles on Azure, especially the difference between training and inference. Training is when you use historical data to create a model (learn parameters). Inference is when you use the trained model to generate predictions on new data (score).

Expect exam scenarios like: “A company wants to build a model to predict churn. Which step happens during training?” Training involves selecting an algorithm, fitting it to labeled data, and tuning parameters. Inference involves deploying the model and sending new customer records to get churn probabilities.

Core ML vocabulary is heavily tested: data (the dataset), features (input variables, like tenure or monthly spend), and labels (the target you want to predict, like churn yes/no). In supervised learning, labels exist during training; in inference, labels are typically unknown (that’s what you’re predicting). The model lifecycle includes data ingestion, feature engineering, training, evaluation, deployment, monitoring, and retraining.

Exam Tip: If the prompt mentions “build,” “fit,” “learn,” “train,” “evaluate,” or “tune,” that’s training. If it mentions “score,” “predict,” “classify new records,” “deploy,” or “real-time endpoint,” that’s inference.

Common trap: confusing “training a model” with “using a model.” A deployed web service that returns a prediction is doing inference—even if it’s running in Azure ML. Another trap is assuming every ML model needs continuous learning. Many production systems retrain periodically (weekly/monthly) rather than learning on-the-fly.

Section 3.2: Learning types: classification, regression, clustering; sample business mappings

AI-900 expects you to map business problems to learning types. The three headline types you must know are classification, regression, and clustering. Classification predicts a category (often yes/no or among multiple classes). Regression predicts a numeric value. Clustering groups similar items when there are no labels.

Classification examples: fraud detection (fraud/not fraud), email routing (sales/support/billing), disease detection (positive/negative), sentiment (positive/neutral/negative). Regression examples: forecasting demand (number of units), predicting house price (dollars), estimating time-to-failure (hours). Clustering examples: customer segmentation, grouping products by similarity, or identifying natural groupings of behaviors.

Unsupervised learning is most often tested through clustering. If the scenario says “we don’t know the groups ahead of time,” “discover segments,” or “find patterns without labels,” think clustering. Conversely, if the scenario includes a known outcome column (churn, cost, category), it’s supervised learning.

Exam Tip: Watch for the label type. If the target is a word/category, it’s classification. If it’s a number (price, temperature, revenue), it’s regression. If there is no target column and you want groups, it’s clustering.

Common trap: confusing “binary classification” with “regression because it outputs a probability.” Many classification models output probabilities (0–1), but the task is still classification because the label is categorical.

Section 3.3: Model training workflow: split data, train, validate, test; overfitting and underfitting

In exam terms, the model training workflow is about preventing “good-on-paper, bad-in-real-life” models. A typical workflow is: split data into training and test sets, and often use a validation set (or cross-validation) during model selection and tuning. Training data teaches the model; validation helps tune choices; the test set gives an unbiased final estimate of performance.

AI-900 questions may describe a team getting excellent results during training but poor results after deployment. That points to overfitting: the model learned noise and specifics of the training data rather than general patterns. Underfitting is the opposite: the model is too simple or not trained well enough to capture the signal, so it performs poorly even on training data.

Exam Tip: If performance is high on training and low on test/new data, think overfitting. If performance is low everywhere, think underfitting (or insufficient features/data quality issues).

Another commonly tested concept is data leakage. Leakage happens when information from the future or from the label “sneaks” into features (for example, using “account closed date” to predict churn). That can inflate validation accuracy but fails in production because those signals are not available at inference time.

Practical answer selection strategy: when asked how to improve generalization, look for options like “collect more representative data,” “reduce model complexity,” “regularize,” “use cross-validation,” or “remove leaky features.” Avoid choices that merely “increase training time” unless paired with a reason (more data or better tuning).

Section 3.4: Evaluation metrics: accuracy, precision/recall, F1, confusion matrix, MAE/MSE/RMSE, silhouette score (conceptual)

The exam tests metric selection and interpretation at a conceptual level. For classification, you should know accuracy, precision, recall, F1, and the confusion matrix. A confusion matrix counts true positives, true negatives, false positives, and false negatives—useful for explaining tradeoffs.

Accuracy is the fraction of correct predictions, but it can be misleading with imbalanced classes (for example, 99% “not fraud” accuracy by predicting “not fraud” always). Precision answers: “When the model predicts positive, how often is it right?” Recall answers: “Of all actual positives, how many did we catch?” F1 balances precision and recall.

Exam Tip: If false positives are costly (blocking good customers), prioritize precision. If false negatives are costly (missing fraud/cancer), prioritize recall. If you need a balance, consider F1.

For regression, know MAE (mean absolute error), MSE (mean squared error), and RMSE (root mean squared error). MAE is easier to interpret in the original units. MSE and RMSE penalize large errors more strongly; RMSE is in original units, while MSE is squared units.

For clustering, AI-900 typically only requires a conceptual understanding of silhouette score: it measures how well points fit within their cluster compared to other clusters. Higher silhouette generally means better separated, more coherent clusters.

Common trap: picking accuracy by default. When the scenario mentions “rare events,” “imbalanced,” “fraud,” “fault detection,” or “medical screening,” accuracy alone is usually the wrong metric choice.

Section 3.5: Azure ML capabilities overview: workspaces, AutoML, designer, and responsible ML concepts

This objective connects ML fundamentals to Azure services. Azure Machine Learning (Azure ML) is the primary platform for building, training, and deploying ML models. The exam expects you to recognize key components rather than implement them.

An Azure ML workspace is the central resource that organizes assets such as datasets, compute targets, experiments, models, endpoints, and monitoring artifacts. If a question asks where you manage models, runs, and deployments, “workspace” is often the best match.

AutoML (Automated Machine Learning) helps you quickly train and compare models by automating algorithm selection and hyperparameter tuning. AutoML is commonly positioned in exam items as the fastest way to create a baseline model when you know the target column and task type (classification/regression) but don’t want to hand-design pipelines.

Azure ML designer is the low-code drag-and-drop interface to build training pipelines visually. If a scenario says a team prefers a graphical interface and minimal coding, designer is the cue.

Responsible ML concepts appear across AI-900. For ML specifically, think fairness, reliability, privacy/security, inclusiveness, transparency, and accountability. The test often frames these as risk controls: checking for biased training data, explaining model behavior to stakeholders, and monitoring drift after deployment.

Exam Tip: If the prompt emphasizes “no/low code,” choose designer or AutoML. If it emphasizes “manage experiments/models/endpoints in one place,” choose workspace. If it emphasizes “fairness/interpretability,” choose responsible AI practices rather than a specific algorithm.

Section 3.6: Exam-style practice: metric selection, learning-type identification, and lifecycle questions

AI-900 questions in this domain are typically short scenarios with one key clue. Your job is to match the clue to (1) learning type, (2) lifecycle stage, or (3) evaluation metric. Build a habit of scanning for the “target” and the “cost of mistakes.”

For learning-type identification, locate the outcome: a category implies classification; a number implies regression; “discover groups” implies clustering. If the question includes “labeled historical examples,” it’s supervised. If it says “no labels,” it’s unsupervised.

For lifecycle questions, identify whether the activity is training/evaluation (experiments, tuning, comparing metrics) or inference (deploying an endpoint, scoring new data, real-time predictions). Words like “generalize,” “test set,” or “cross-validation” anchor you in training/evaluation. Words like “API,” “endpoint,” or “production scoring” anchor you in inference.

For metric selection, look for imbalanced data and asymmetric costs. Fraud and rare defect detection typically push you toward recall (don’t miss positives) or precision (don’t flag too many false alarms), or F1 as a compromise. For regression, interpretability often favors MAE; sensitivity to outliers often points to RMSE/MSE.

Exam Tip: When two answers both seem plausible, ask: “What is the label type?” and “Which error is worse—false positive or false negative?” Those two checks eliminate many distractors.

Common trap: mixing up clustering evaluation with classification metrics. Accuracy, precision, and recall require labels; silhouette score is for clustering where labels are absent. Another trap is assuming AutoML is only for experts—on AI-900 it’s presented as an accessibility feature for quickly building models with minimal code.

Chapter milestones
  • Understand core ML concepts: data, features, labels, and model lifecycle
  • Distinguish supervised vs unsupervised learning and common algorithms
  • Interpret evaluation metrics and avoid common pitfalls
  • Domain practice set: ML principles on Azure
Chapter quiz

1. A retail company wants to predict whether an online order will be returned (Yes/No) based on order value, shipping method, and customer tenure. Which machine learning approach best fits this requirement?

Show answer
Correct answer: Supervised learning classification
This is a supervised learning problem because historical outcomes (returned: Yes/No) can be used as labels. The goal is to predict a categorical class, so classification is appropriate. Regression is for predicting a numeric value (for example, return amount), not a Yes/No outcome. Clustering is unsupervised and groups records without labels; it doesn’t directly produce a Yes/No prediction tied to known outcomes.

2. You are reviewing a dataset used to train a model that predicts house prices. Which statement correctly identifies features and labels in this scenario?

Show answer
Correct answer: The sale price is the label; the number of bedrooms and square footage are features
In supervised learning, the label is the value you want to predict (sale price), and features are the input variables (bedrooms, square footage, location, etc.). Treating sale price as a feature and the inputs as labels reverses the model’s purpose. Saying all columns are features is incorrect for supervised training because a target label is required to learn a mapping and evaluate performance.

3. A healthcare provider builds a model to detect a rare condition. Missing a true case is costly, and the data is highly imbalanced (very few positive cases). Which metric is generally more appropriate to prioritize during evaluation?

Show answer
Correct answer: F1 score
With imbalanced classes and high cost of false negatives, accuracy can be misleading (a model can be highly accurate by predicting the majority class). F1 score balances precision and recall, making it more informative when positives are rare and you care about capturing them while managing false alarms. R-squared is used for regression scenarios and does not apply to a binary classification task.

4. A team wants to segment customers into groups based on purchasing behavior without any pre-defined categories. Which type of machine learning is most suitable?

Show answer
Correct answer: Unsupervised learning
Customer segmentation without known categories uses unsupervised learning, commonly clustering, because there are no labels to train against. Supervised learning requires labeled outcomes (for example, churn: Yes/No). Reinforcement learning is used for sequential decision-making with rewards (for example, optimizing actions over time), not for static grouping of records.

5. A company has historical data and wants to build and deploy a model in Azure with minimal coding. They also want Azure to automatically try different algorithms and feature transformations to find the best model. Which Azure capability best matches this need?

Show answer
Correct answer: Azure Machine Learning Automated ML (AutoML)
Azure Machine Learning AutoML is designed to automate model selection and tuning (including trying algorithms and transformations) for common ML tasks, aligning with a low-code approach for building models. Cognitive Services vision APIs are prebuilt AI services for specific workloads (like image analysis) rather than training a custom model from tabular business data. Azure Monitor is for operational monitoring and logging; it does not train or select machine learning models.

Chapter 4: Computer Vision Workloads on Azure

Computer vision questions on AI-900 are mostly about recognizing the workload (what the business wants the AI to do) and matching it to the right Azure capability—without getting pulled into unnecessary implementation details. Expect exam prompts that describe everyday scenarios (retail shelves, invoices, ID badges, manufacturing defects, photo moderation) and ask what type of vision task it is (classification vs object detection vs OCR), what outputs you should expect (tags, captions, bounding boxes, confidence scores, extracted text), and what constraints/responsible considerations apply.

At a high level, computer vision “inputs” are images (JPEG/PNG), image URLs, and sometimes video frames. The “outputs” are structured JSON-like results containing labels/tags, natural-language descriptions, object locations, and recognized text. The exam tests whether you can interpret those outputs conceptually: what a bounding box means, what a confidence score implies, and how thresholds affect results.

Exam Tip: When the scenario says “find where the item is in the image,” that is object detection (location + label). When it says “what is in the image,” that is image analysis/classification (labels/tags without precise location). When it says “read the text,” that is OCR. These three words—where, what, read—often unlock the correct answer.

  • Classification / image analysis: identify content (e.g., “this is a dog,” “outdoor,” “car”).
  • Object detection: identify and localize content (e.g., “car at these coordinates”).
  • OCR: extract printed/handwritten text from images or documents.

This chapter builds your service-selection instincts, explains typical vision outputs and constraints, and reinforces responsible patterns—especially around people-focused use cases.

Practice note for Identify core vision tasks and the right Azure service fit: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand image analysis, OCR, and video/face scenarios at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Know typical inputs/outputs, constraints, and responsible considerations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: computer vision workloads: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify core vision tasks and the right Azure service fit: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand image analysis, OCR, and video/face scenarios at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Know typical inputs/outputs, constraints, and responsible considerations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: computer vision workloads: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify core vision tasks and the right Azure service fit: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Official objective—Computer vision workloads on Azure: classification, object detection, OCR

AI-900 expects you to distinguish the three core vision workloads that appear repeatedly: classification (often phrased as image analysis), object detection, and optical character recognition (OCR). The most common exam trap is confusing “classification” with “object detection.” Both can identify objects, but only detection provides locations (bounding boxes). Another trap is thinking OCR is only for PDFs; in exams it can be a photo of a sign, a screenshot, or a scanned form.

Classification / image analysis answers: “What is this image about?” Outputs tend to be tags (keywords), categories, and a caption/description. Use this when you need to label or describe whole images, support search, or route images to the right workflow. Object detection answers: “What objects are present and where?” Outputs include object names plus bounding box coordinates. Use it for counting items, finding defects, locating products on a shelf, or triggering actions based on where something appears in a frame.

OCR answers: “What text is visible?” It returns extracted text plus positions. OCR scenarios include digitizing receipts, reading serial numbers, extracting sign text for accessibility, and pulling text from screenshots. On AI-900, the right answer is rarely about training custom models; it’s about selecting a prebuilt capability in Azure AI Vision (and, for documents, knowing that document-focused extraction is often handled by a document-intelligence style capability rather than generic image tagging).

Exam Tip: Watch for verbs. “Detect,” “locate,” “count,” “track” → object detection. “Describe,” “tag,” “categorize” → image analysis. “Extract,” “read,” “recognize text” → OCR.

Section 4.2: Azure AI Vision overview: image analysis capabilities and common outputs

Azure AI Vision (often referred to on the exam as “Azure AI Vision” or “Vision services”) provides prebuilt capabilities for analyzing images. The exam focuses on what you can get back from an image, not SDK calls. Typical image analysis outputs include: tags (keywords like “person,” “outdoor,” “vehicle”), captions (a short natural-language description), objects (identified items, sometimes with location), and content moderation signals (depending on the feature set referenced in the prompt).

Interpretation matters. A tag list is useful for search indexing and filtering (“show me all images containing ‘dog’”). A caption is useful for accessibility (“describe this image to a user”). Object results are useful when the system must act on a specific part of the image (“blur faces,” “highlight products”). Many exam questions simply provide a scenario and ask what output is appropriate, so align the output to the business need: search needs tags, accessibility needs captions, automation needs detection with coordinates.

Common exam trap: assuming “image classification” implies a custom-trained model. In AI-900, most prompts point to prebuilt image analysis unless they explicitly say you have labeled images and need to train for a niche category. If the prompt includes unusual domain objects (specialized parts, proprietary product types) and mentions providing labeled examples, that’s when you start thinking beyond generic image analysis—otherwise, default to prebuilt Vision capabilities.

Exam Tip: If the question mentions “confidence” with tags/captions/objects, treat it as a probability-like score. Higher confidence means the service is more certain. The exam may test whether you should set a threshold (e.g., only accept tags above 0.8) to reduce false positives.

Section 4.3: OCR and document concepts: text extraction, layout ideas, and when to use which option

OCR is about converting visual text into machine-readable text. For AI-900, you should recognize two levels of need: (1) simple text extraction from an image (signs, labels, screenshots), and (2) document-oriented extraction where layout and structure matter (forms, invoices, receipts, multi-page documents). Both start with OCR, but the second typically requires understanding layout elements such as lines, words, and their positions, and sometimes key-value pairing (e.g., “Invoice Number: 12345”).

When the scenario says “extract all text from this photo,” basic OCR is sufficient. When it says “extract fields from invoices” or “ingest forms and preserve structure,” treat it as a document concept: the system cares about where the text is and what it means in context. The exam often checks whether you can articulate that OCR may return not only text but also coordinates (bounding polygons/boxes for words and lines), which enables highlighting text in a UI or mapping text into columns.

Common exam trap: picking image tagging/classification when the business asks to “find the account number on a statement.” Tags won’t give you the string; OCR will. Another trap is assuming OCR always produces perfect results. Real-world constraints (blur, handwriting, angles, lighting) reduce accuracy; this leads into responsible patterns such as human review or validation rules for high-impact data entry.

Exam Tip: If the prompt mentions “layout,” “tables,” “forms,” “key-value pairs,” or “fields,” think document-style extraction rather than general image description. If it only mentions “text in an image,” OCR is the anchor capability.

Section 4.4: Spatial and detection scenarios: bounding boxes, tags, confidence scores, thresholds

Spatial understanding is what differentiates object detection from image-level labeling. In detection results, each object typically includes a label/name, a bounding box (coordinates defining the object’s location in the image), and a confidence score. On the exam, bounding boxes are usually discussed conceptually: they let you draw rectangles around detected items, count them, crop them, or trigger business logic (“if a person is detected in a restricted area, alert security”).

Confidence scores are central to how you interpret results. A high-confidence detection is more reliable; a low-confidence detection may be noise. That’s why systems commonly use thresholds: only accept detections above a certain confidence. The exam may test the trade-off: raising a threshold reduces false positives but can increase false negatives (missing real objects). Lowering a threshold captures more candidates but increases the need for secondary checks.

Tags vs bounding boxes is a frequent decision point. If you only need to know whether an image contains a bicycle, tags might be enough. If you need to know where the bicycle is to blur it, count it, or ensure it’s present in a specific region, you need detection (bounding boxes). Similarly, OCR returns location information for text, enabling “click to highlight” or redaction workflows.

Exam Tip: When a question includes UI actions like “draw a rectangle,” “highlight,” “crop,” “mask,” or “redact,” that implies spatial coordinates are required—choose an option that returns bounding boxes/polygons, not just labels.

Section 4.5: Responsible vision: bias, privacy, and human review patterns for sensitive scenarios

AI-900 explicitly expects responsible AI awareness, including for vision. Vision workloads can introduce privacy and bias risks because images may contain people, locations, IDs, or other sensitive information. The exam is not asking you to implement governance frameworks, but it does test whether you can recognize when extra safeguards are needed.

Privacy: If images include faces, license plates, ID cards, medical images, or children, you should consider data minimization (collect only what you need), secure storage, access controls, and retention policies. Also consider whether you can process images without storing them long-term, and whether you can redact sensitive regions using detection/OCR location outputs.

Bias and fairness: People-centered vision (especially face-related analysis) can perform unevenly across demographic groups depending on training data. The exam may frame this as a risk that requires testing across representative datasets and monitoring performance. Even when the service is prebuilt, you are responsible for validating it in your context.

Human review patterns: In high-impact scenarios (identity verification, safety decisions, legal/HR outcomes), use human-in-the-loop review—particularly for low-confidence results or edge cases. Set thresholds, route uncertain items to manual validation, and log decisions for auditing.

Exam Tip: If a scenario mentions surveillance, identification, hiring, law enforcement, or processing personal data, look for answers that include consent, privacy controls, human review, and transparency. A common trap is choosing “fully automate decisions” when the scenario is sensitive or high impact.

Section 4.6: Exam-style practice: service selection and interpreting vision results

On AI-900, the highest-value skill is mapping a scenario statement to (1) the vision task and (2) the expected output shape. Practice reading prompts and underlining the deliverable: “describe,” “find,” “count,” “read,” “extract fields,” “highlight,” “redact.” Then choose the Azure AI Vision capability (image analysis, object detection, OCR) that naturally returns the needed artifact (tags/caption, bounding boxes, extracted text with positions).

Service-selection cues you should recognize quickly: if the scenario is about cataloging photos or improving search, image analysis with tags/captions is the fit. If it’s about inventory counting, safety zones, defect localization, or where an item appears, object detection is the fit. If it’s about digitizing printed/handwritten text, use OCR; if it adds forms/invoices/receipts and structured fields, treat it as document-style extraction layered on OCR concepts.

Interpreting results is also tested. If you see multiple objects with confidence scores, the correct operational answer often involves applying a threshold and optionally a review step. If the prompt mentions errors like false detections, the likely fix is raising the threshold or adding validation logic—not “train a new model,” unless the prompt explicitly mentions a custom domain and labeled training data.

Exam Tip: When two answers both “sound plausible,” pick the one whose output matches the scenario’s required action. Needing coordinates → detection/OCR with location. Needing keywords → tags. Needing readable strings → OCR, not classification.

Chapter milestones
  • Identify core vision tasks and the right Azure service fit
  • Understand image analysis, OCR, and video/face scenarios at a high level
  • Know typical inputs/outputs, constraints, and responsible considerations
  • Domain practice set: computer vision workloads
Chapter quiz

1. A retailer wants to build an app that identifies products on a shelf and returns each product’s location in the image so staff can see which items are out of place. Which computer vision task best fits this requirement?

Show answer
Correct answer: Object detection
The key requirement is to find where each product is located (coordinates), which is object detection (label + bounding box). Image classification/image analysis answers what is in the image (tags/captions) but does not provide precise locations. OCR is used to read text from images, which is not the goal here.

2. A company receives scanned invoices as JPEG files and needs to extract the invoice number, vendor name, and total amount from the images. Which type of computer vision workload should you use?

Show answer
Correct answer: OCR
Extracting printed (and sometimes handwritten) text from scanned documents is an OCR workload. Object detection would locate objects (for example, a logo or stamp) but does not inherently return the text content. Image classification returns tags/captions about what the image contains (for example, “document,” “paper”) rather than the actual characters and words.

3. You call a vision service and receive results that include labels like "car" and "person" along with coordinates for rectangles surrounding each item and a confidence score per item. What does the rectangle represent?

Show answer
Correct answer: A bounding box that indicates the detected object’s location
Rectangles with coordinates in vision outputs represent bounding boxes, which are produced by object detection to localize items in an image. A classification label applies to the whole image and does not require coordinates. OCR can return regions/lines/words, but the scenario explicitly ties rectangles to detected objects like "car" and "person," which aligns with object detection outputs in AI-900 expectations.

4. A social media company wants to automatically describe user-uploaded photos with tags such as "beach," "sunset," and "outdoor" to improve search. The company does not need to know where objects appear in the photo. Which workload is most appropriate?

Show answer
Correct answer: Image classification (image analysis)
Generating tags/captions about what is in an image without locating items is image classification/image analysis. Object detection is used when the scenario requires where an item is (bounding boxes). OCR is for reading text from images and is unrelated to generating general content tags like "sunset" or "outdoor."

5. A manufacturing team uses object detection to identify defects on parts. They plan to increase the confidence threshold so only high-confidence detections are returned. What is a likely impact of increasing the confidence threshold?

Show answer
Correct answer: Fewer detections will be returned, potentially missing some real defects
Raising the confidence threshold filters out lower-confidence results, typically reducing the number of returned detections and potentially increasing false negatives (missed true defects). Returning more detections (including more false positives) is more likely when the threshold is lowered. Threshold changes do not change the type of workload (object detection vs OCR); they only affect which results are included.

Chapter 5: NLP and Generative AI Workloads on Azure

This chapter maps directly to the AI-900 objectives around natural language processing (NLP), speech, and generative AI workloads. The exam expects you to recognize common business scenarios, match them to the right Azure service family, and use correct workload vocabulary (for example: sentiment analysis vs. entity recognition, speech-to-text vs. text-to-speech, and prompts vs. embeddings). Non-technical candidates often lose points not because the concepts are hard, but because the question wording is subtle: “extract information” typically implies entities/key phrases; “classify” implies categories; “summarize” implies condensation; “chatbot” implies conversational experiences; and “generate” implies a foundation model workload.

In this chapter, you’ll practice choosing the right NLP task (sentiment, key phrases, entities, summarization), understanding speech scenarios (transcription, synthesis, translation), and explaining generative AI basics and Azure OpenAI use cases safely. Your goal is to build “pattern recognition” for exam prompts: identify the workload first, then select the Azure capability that naturally supports that workload.

  • NLP: analyze, categorize, and extract meaning from text
  • Speech: convert audio to text, text to audio, or translate across languages
  • Generative AI: create new text/code/images from prompts using foundation models

As you read, focus on (1) the scenario keywords, (2) the expected output, and (3) the service family that matches the task. Those three steps consistently lead you to the correct answer on AI-900.

Practice note for Choose the right NLP task: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand speech scenarios: speech-to-text, text-to-speech, translation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI basics and Azure OpenAI use cases safely: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: NLP + generative AI workloads: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right NLP task: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand speech scenarios: speech-to-text, text-to-speech, translation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI basics and Azure OpenAI use cases safely: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: NLP + generative AI workloads: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right NLP task: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: Official objective—NLP workloads on Azure: analyze text and build conversational experiences

Section 5.1: Official objective—NLP workloads on Azure: analyze text and build conversational experiences

On AI-900, an “NLP workload” usually means one of two big buckets: analyze text (understand what’s in text) or build conversational experiences (interact with users via chat/voice). You’re not expected to code, but you are expected to identify which capability fits the business need and what the output looks like.

Text analysis questions often describe inputs such as emails, support tickets, survey responses, social posts, or documents. The exam then asks what you should do: detect opinion, extract topics, find names/locations, or summarize. Conversational experience questions describe a virtual agent, FAQ bot, or customer support assistant that needs to understand user utterances and respond appropriately.

How to identify the correct NLP task: look for the “verb” in the scenario. “Determine whether feedback is positive or negative” → sentiment. “Pull out product names and people” → entities. “List the main topics customers mention” → key phrases. “Create a short version of a long article” → summarization.

Common exam trap: mixing up key phrases vs. entities. Key phrases are important terms (often noun phrases) that represent topics, while entities are specific types like person, organization, location, date/time, or custom domain entities. If the question mentions “types” (person, address) or “identify proper nouns,” it’s entity recognition. If it says “main talking points,” it’s key phrase extraction.

Exam Tip: When a scenario says “classify messages into categories like billing, technical, or sales,” that is text classification, not entity extraction. Classification outputs a label (and often a confidence score). Extraction outputs spans of text.

For conversational experiences, the exam focus is conceptual: you need NLP to understand intent and language, and you often integrate speech if the experience is voice-based. Don’t overcomplicate: first decide if the user is interacting (conversation) or you are analyzing stored text (batch analysis). That decision alone eliminates many wrong options.

Section 5.2: Azure AI Language overview: classification, entity recognition, sentiment, summarization (conceptual)

Section 5.2: Azure AI Language overview: classification, entity recognition, sentiment, summarization (conceptual)

Azure AI Language is the core service family you map to “analyze text” tasks on AI-900. The exam does not require endpoint details, but you must know the capability names and when to use each one. Think in terms of what the model returns: labels, extracted text spans, a sentiment score, or a summary.

Text classification assigns categories to text. Typical scenarios: route support tickets, detect spam vs. legitimate messages, tag customer inquiries by department. The output is usually a class label (for example, “Billing”) plus a confidence. A frequent distractor is sentiment analysis—sentiment is not “classification into business categories,” it’s opinion polarity.

Entity recognition (including PII detection in some contexts) extracts structured information from unstructured text: names, organizations, locations, dates, product IDs, and domain-specific entities. Use entity recognition when the question mentions “extract,” “identify,” “pull out,” “find,” or building downstream processes like populating a database.

Sentiment analysis determines opinion/attitude, commonly returned as positive/neutral/negative and/or scores. Watch for scenario keywords: “customer satisfaction,” “brand perception,” “tone,” “how customers feel.”

Summarization condenses content (for example, long emails, meetings, or reports) into shorter text. If the scenario says “generate a short abstract” or “highlight the main points,” summarization is your match. If it says “extract topics,” that’s key phrases, not summarization.

Exam Tip: Many questions include multiple correct-sounding NLP tasks. Always anchor on the requested output. If they want “a list of key topics,” select key phrase extraction. If they want “a short paragraph overview,” select summarization. If they want “a rating of positivity,” select sentiment.

In real projects, these tasks can be combined (for example, summarize and extract key phrases), but the exam typically wants the primary capability that directly meets the requirement. Choose the most direct mapping and avoid “kitchen sink” thinking.

Section 5.3: Azure AI Speech overview: transcription, synthesis, translation; typical business scenarios

Section 5.3: Azure AI Speech overview: transcription, synthesis, translation; typical business scenarios

Azure AI Speech maps to workloads where the input or output is audio. The AI-900 exam commonly tests three ideas: speech-to-text (transcription), text-to-speech (synthesis), and speech translation. The trick is to identify the direction of conversion and whether translation is required.

Speech-to-text (STT) converts spoken audio into written text. Scenarios include call center transcription, meeting notes, voice dictation, and indexing audio archives for search. If the question mentions “transcribe,” “caption,” “convert audio recordings to text,” or “generate subtitles,” you’re in STT territory.

Text-to-speech (TTS) converts text into natural-sounding audio. Scenarios include reading news articles aloud, accessibility features, IVR systems that “speak” dynamic content, or voice responses in an assistant. If the prompt says “create an audio voice from text” or “read out responses,” that’s TTS.

Speech translation converts spoken language to another language, often producing text and/or synthesized speech in the target language. This appears in scenarios such as multilingual support desks, real-time translation for events, or travel apps. A common trap is choosing “Translator” for any language problem. If the input is audio, prioritize Speech translation; if the input is text, translation is a text NLP capability.

Exam Tip: If you see “real-time captions in another language,” you need two operations: speech-to-text plus translation. The correct exam choice is usually the service that explicitly supports speech translation rather than stitching separate tools in your head.

Business scenario pattern recognition helps: call centers → STT and analytics; voice assistants → STT + language understanding + TTS; accessibility → TTS; multilingual meetings → speech translation. When you practice, always ask: “What is the input type (audio/text) and what is the required output type (audio/text/translated)?”

Section 5.4: Official objective—Generative AI workloads on Azure: foundation models and prompt concepts

Section 5.4: Official objective—Generative AI workloads on Azure: foundation models and prompt concepts

Generative AI on AI-900 is about recognizing scenarios where the system creates new content rather than merely classifying or extracting. The exam will use terms like generate, draft, compose, rewrite, summarize in a new style, or produce code. These point to foundation models: large, general-purpose models trained on broad datasets and then adapted to tasks via prompting and/or fine-tuning.

A prompt is the input instruction and context you provide to a model. Prompts can include role guidance (“You are a support agent”), constraints (“use bullet points”), and examples (“few-shot” prompting). The output is generated probabilistically, which means the same prompt can produce slightly different answers depending on settings.

Common exam trap: confusing summarization as a classic NLP feature vs. generative AI summarization. On the exam, if the scenario emphasizes “generate a new short version” or “draft a response” in natural language, it often hints at generative AI. If it emphasizes “analyze text and return a summary feature,” it can still be Azure AI Language summarization. Use the clue: is the task framed as analysis of existing text or as content creation for a user/workflow?

Exam Tip: Generative AI questions usually mention one of these: chat experiences, prompt engineering, foundation models, embeddings, or Azure OpenAI. If you see those keywords, you’re likely in the generative AI objective, not classic NLP.

Also expect the exam to test safe and appropriate usage at a conceptual level: models can be powerful but can make mistakes (“hallucinate”), may need grounding in trusted data, and require guardrails. Those responsible AI concepts are covered more deeply in Section 5.6, but you should already be thinking: “What are the risks, and how does Azure help mitigate them?”

Section 5.5: Azure OpenAI basics: prompts, completions/chat, embeddings, common enterprise use cases

Section 5.5: Azure OpenAI basics: prompts, completions/chat, embeddings, common enterprise use cases

Azure OpenAI is Microsoft’s Azure service for using OpenAI models with enterprise-oriented controls. For AI-900, you need to know the foundational building blocks: prompts, chat/completions, and embeddings, plus typical use cases.

Chat / completions refer to generating text responses from a prompt. Chat is optimized for multi-turn conversations (messages with roles like system/user/assistant). “Completions” is the general idea of continuing or producing text. Scenarios: drafting emails, writing product descriptions, summarizing and rewriting content, generating FAQs, creating a customer support assistant, or helping internal employees find answers.

Embeddings are numeric vector representations of text (and sometimes other content types) that capture semantic meaning. You use embeddings for “find similar,” semantic search, clustering, recommendations, and retrieval-augmented generation (RAG) patterns where you fetch relevant documents and then generate an answer grounded in them.

Common exam trap: picking “chat/completions” for a search scenario. If the requirement is “search for similar documents,” “semantic similarity,” or “rank results by meaning,” embeddings are the better conceptual match. Generation produces text; embeddings enable similarity and retrieval.

Exam Tip: When you see “vector,” “semantic search,” “similarity,” or “retrieve relevant passages,” think embeddings. When you see “draft,” “write,” “generate,” or “conversational response,” think chat/completions.

Enterprise use cases to memorize (because they appear repeatedly): employee Q&A over policies, customer support copilots, summarizing call notes, document drafting, and classification/triage assistance. The exam typically rewards choosing the simplest statement: Azure OpenAI is for generative tasks; Azure AI Language is for classic NLP analysis tasks; Azure AI Speech is for audio scenarios.

Section 5.6: Responsible generative AI: hallucinations, grounding, safety filters, and data privacy

Section 5.6: Responsible generative AI: hallucinations, grounding, safety filters, and data privacy

AI-900 includes responsible AI expectations, and generative AI raises specific risks. A top test concept is hallucinations: the model may produce fluent but incorrect content. Another is grounding: connecting the model’s response to trusted enterprise data (for example, internal documents) so answers are based on verifiable sources rather than “best guesses.” If a scenario says “ensure responses are based on company policy documents,” the right idea is grounding (often implemented via retrieval before generation).

Safety filters and content moderation are also emphasized. Azure provides mechanisms to help detect and reduce harmful content (for example, hate, sexual, violence, self-harm). On the exam, if the scenario mentions “prevent inappropriate responses” or “content policy enforcement,” look for choices referencing safety systems, content filtering, or responsible AI guardrails—not model accuracy features.

Data privacy is a common non-technical concern: organizations want to know whether prompts and outputs are handled securely, and whether sensitive data is protected. Exam questions may frame this as “avoid exposing confidential information” or “protect customer data.” Your best conceptual response is to apply least privilege, data governance, and avoid placing secrets in prompts; also prefer architectures where you control data sources and access.

Common exam trap: assuming the model is a database of truth. Generative models predict likely text; they do not inherently verify facts. When the question asks how to improve trustworthiness, the best answer is usually “ground the response in trusted data” and “provide citations,” not “increase temperature” or “use a larger model.”

Exam Tip: If the question asks how to reduce hallucinations, pick grounding/retrieval and clear system instructions over “more training data” or “more tokens.” If it asks how to reduce harmful outputs, pick safety filters/content moderation and policy-based controls.

As you work domain practice for NLP + generative AI workloads, apply a final checklist: What is the input (text/audio)? What is the output (labels/extractions/summary/new content/audio)? Is the main goal analysis or generation? And what responsible AI risk must be addressed (accuracy, safety, privacy)? Those four questions align closely with how AI-900 scenarios are written and how correct answers are distinguished from distractors.

Chapter milestones
  • Choose the right NLP task: sentiment, key phrases, entities, summarization
  • Understand speech scenarios: speech-to-text, text-to-speech, translation
  • Explain generative AI basics and Azure OpenAI use cases safely
  • Domain practice set: NLP + generative AI workloads
Chapter quiz

1. A retail company wants to process thousands of product reviews and determine whether customers are expressing positive, negative, or neutral opinions. Which NLP task should you use?

Show answer
Correct answer: Sentiment analysis
Sentiment analysis is designed to classify text by opinion polarity (positive/negative/neutral). Key phrase extraction identifies important terms but does not score opinion. Entity recognition finds named items (people, locations, organizations) and is not intended to determine sentiment.

2. A support team receives long email threads and wants an automated feature that condenses each thread into a short overview for agents. Which NLP capability best fits this requirement?

Show answer
Correct answer: Text summarization
Text summarization produces a condensed version of longer content, matching the requirement to create an overview. Entity recognition extracts named entities (for example, product names or customer names) but does not condense content. Language detection identifies the language of the text and does not produce a summary.

3. A medical transcription service needs to convert recorded doctor-patient conversations into written text for review. Which speech scenario is required?

Show answer
Correct answer: Speech-to-text
Speech-to-text transcribes audio into text, which is the stated goal. Text-to-speech converts written text into synthesized audio, the opposite direction. Speech translation focuses on converting speech in one language into another language (often as text and/or audio), which is not required if the goal is transcription.

4. A travel company wants a phone-based system where a caller speaks Spanish and the agent receives the content in English in near real time. Which speech workload best matches this scenario?

Show answer
Correct answer: Speech translation
Speech translation is intended to translate spoken language into another language, aligning with Spanish-to-English during a call. Text-to-speech generates audio from text and does not translate incoming speech. Key phrase extraction is an NLP text analysis task and does not perform real-time spoken translation.

5. A company wants to build an internal assistant that drafts policy FAQ responses from user prompts. They also want to reduce the chance the assistant outputs harmful or inappropriate content. Which Azure capability is the best fit?

Show answer
Correct answer: Azure OpenAI (generative AI) with safety controls such as content filtering
Drafting natural-language responses from prompts is a generative AI workload, which aligns with Azure OpenAI; using safety features (for example, content filtering and responsible AI practices) addresses the requirement to reduce harmful output. Key phrase extraction only pulls important terms and cannot draft full answers. Text-to-speech converts text to audio and does not generate the policy responses.

Chapter 6: Full Mock Exam and Final Review

This final chapter is where you convert “I’ve read it” into “I can pass it.” AI-900 is designed for non-technical professionals, but the exam still tests precision: choosing the right Azure AI service for a scenario, recognizing ML fundamentals (training vs. inference, classification vs. regression), and applying responsible AI concepts. Your goal is not memorization of product brochures; it’s fast scenario-to-solution mapping aligned to the exam objectives.

You will complete two mock exam passes (Part 1 and Part 2), then perform weak spot analysis, and finish with a rapid final review sprint. The outcome: consistent, repeatable decision-making under time pressure. Throughout, use an objective-first mindset: each item on the exam is written to validate a specific skill statement (workloads, Azure AI services, ML basics/AutoML, vision, NLP/speech, and generative AI with responsible use).

As you work, keep a running “confusion list” of terms that look similar (for example, Azure AI Vision vs. Azure AI Document Intelligence; Azure AI Language vs. Speech; Azure OpenAI vs. generic “chatbot” language). These are frequent trap areas because options may all sound plausible unless you anchor on the workload definition and the service boundary.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final Review Sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final Review Sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam instructions and timing strategy (AI-900 pacing)

Section 6.1: Mock exam instructions and timing strategy (AI-900 pacing)

Before you start the mock exam, decide how you will manage time and uncertainty. AI-900 questions are typically short, scenario-based, and mixed across domains. The most common failure mode is not lack of knowledge, but losing time debating between two “almost correct” services. Your pacing strategy should prioritize quick wins and systematic elimination of distractors.

Use a two-pass approach. Pass 1: answer what you can in one minute or less per item. If you feel the urge to “research in your head,” mark it and move on. Pass 2: return to marked items with fresh context, using objective mapping (What workload is this? Which service is purpose-built for it? What responsible AI concern applies?).

Exam Tip: In mock conditions, practice making a decision even when you are only 70–80% confident. AI-900 is designed so that the correct option is the best fit, not merely “possible.” Your job is to identify the best match, not every match.

  • Identify keywords first: “detect objects,” “extract text,” “sentiment,” “speech to text,” “predict numeric value,” “generate content,” “evaluate bias,” etc.
  • Translate the scenario into a workload label: computer vision, NLP, speech, ML, or generative AI.
  • Then pick the Azure service family that owns that workload (Vision, Language, Speech, Azure Machine Learning, Azure OpenAI).

Finally, simulate the real environment: no notes, no pausing, and minimal distractions. The goal of this section is to build a consistent process, so your performance is stable on exam day.

Section 6.2: Mock Exam Part 1: mixed-domain questions (exam-style)

Section 6.2: Mock Exam Part 1: mixed-domain questions (exam-style)

Mock Exam Part 1 should feel “wide rather than deep.” Expect rapid switching: one item about responsible AI principles, the next about choosing an Azure AI service, then an ML evaluation concept. Your technique is to anchor on what the question is truly testing: workload identification and correct service selection are the highest-yield skills.

When reviewing Part 1 results, categorize misses into three buckets. Bucket A: you misread the workload (for example, confusing OCR/text extraction with image classification). Bucket B: you knew the workload but chose the wrong service (often Vision vs. Document Intelligence, Language vs. Speech). Bucket C: you understood the service but missed a concept detail (training vs. inference; precision/recall; overfitting; responsible AI principle).

Exam Tip: Watch for “multi-step” scenarios hidden inside short text. If it says “extract text from invoices” that’s document processing (Document Intelligence). If it says “describe what is in the image” that’s Vision (image analysis). If it says “summarize customer calls,” that might be Speech-to-text first, then Language summarization.

Common traps in Part 1 include picking general compute services (VMs, Functions) when the question is clearly about Azure AI services. AI-900 typically wants the managed AI capability, not infrastructure. Another trap: assuming “AutoML” means “no data prep.” AutoML reduces manual model selection and tuning, but you still need representative, well-labeled data and clear evaluation.

Use Part 1 to validate your baseline: can you map scenarios to the correct domain quickly? If not, your weak spot is likely foundational service boundaries, not advanced ML theory.

Section 6.3: Mock Exam Part 2: mixed-domain questions (exam-style)

Section 6.3: Mock Exam Part 2: mixed-domain questions (exam-style)

Mock Exam Part 2 should lean into “edge cases” and the modern AI-900 emphasis: responsible AI and generative AI. You should be comfortable distinguishing predictive ML from generative AI. Predictive ML outputs labels or numbers (classification/regression). Generative AI outputs new content (text, images, code) based on patterns learned from data and guided by prompts.

In this part, the exam often tests whether you can apply responsible AI thinking to realistic scenarios: fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. The correct answer is frequently the one that introduces governance or safeguards rather than “just build it.” For Azure OpenAI scenarios, anticipate questions about content filtering, human-in-the-loop review, grounding (using trusted data sources), and prompt injection awareness at a conceptual level.

Exam Tip: If an option adds monitoring, auditability, or an approval workflow for high-impact decisions, it often aligns with accountability and reliability objectives—especially in HR, lending, healthcare, or public-sector examples.

Part 2 also surfaces ML lifecycle fundamentals: training data vs. test data, model evaluation, and overfitting. A common distractor claims that high training accuracy proves success. On the exam, the best choice usually references validation/testing, generalization, and appropriate metrics. Remember that metric choice depends on business risk: false positives vs. false negatives. In non-technical wording, the question may describe “missing fraud” (false negatives) or “flagging good customers” (false positives).

By the end of Part 2, you should have a clear picture of whether your remaining gaps are conceptual (what the workload is) or operational (how to justify the best option).

Section 6.4: Answer review framework: why the correct option wins (objective-based)

Section 6.4: Answer review framework: why the correct option wins (objective-based)

Weak spot analysis only works if you review answers the right way. Do not focus on “why my choice was wrong” first. Focus on “what objective is being tested” and “what signal in the scenario proves the winning option is best.” This prevents you from learning trivia and instead strengthens your decision rules.

Use a four-step review framework for every missed or guessed item. Step 1: label the domain (AI workloads/responsible AI, ML/AutoML, Vision, NLP/Speech, Generative AI). Step 2: extract the single most important requirement (for example: “extract key-value pairs,” “transcribe audio,” “detect faces,” “classify emails,” “generate a draft response,” “evaluate bias”). Step 3: match to the Azure capability that is purpose-built for that requirement. Step 4: eliminate distractors by stating their mismatch in one sentence.

Exam Tip: Practice writing a one-line justification: “This is document extraction, so Document Intelligence fits; Vision is for general image analysis, not structured invoice fields.” If you cannot write the one-liner, you don’t own the concept yet.

  • Service boundary trap: Language vs. Speech. If the input is audio, Speech services come first; Language services analyze text.
  • Workload trap: OCR vs. image classification. OCR is text extraction; classification assigns categories.
  • Generative vs. predictive trap: “Create a summary” implies generation; “predict churn” implies ML classification.
  • Responsible AI trap: “Accuracy” is not fairness; “Explainability” is not privacy. Map each mitigation to the correct principle.

When you can consistently explain why the correct option wins, your score will stabilize—even when the wording changes.

Section 6.5: Final review by domain: last-pass checklist of exam objectives

Section 6.5: Final review by domain: last-pass checklist of exam objectives

This final review sprint is your “last pass” across the published AI-900 objective areas. Keep it practical: you are verifying recall and decision rules, not re-learning. Use the checklist below and stop when you can answer each prompt confidently in plain language.

  • AI workloads + Responsible AI: Define computer vision, NLP, speech, ML, and generative AI workloads; state the six responsible AI principles; know typical mitigations (human review, monitoring, privacy controls, documentation, fairness evaluation).
  • Machine Learning on Azure: Classification vs. regression vs. clustering; training vs. inference; features vs. labels; train/validate/test purpose; overfitting concept; evaluation metrics at a high level; what AutoML does (model selection/tuning) and what it does not do (replace data quality and objective definition).
  • Computer Vision on Azure: When to use Azure AI Vision for image analysis and detection; when document-centric needs imply Azure AI Document Intelligence; understand OCR as extracting text and layout vs. labeling an image.
  • NLP + Speech on Azure: Azure AI Language for sentiment, key phrases, entity recognition, classification, summarization; Azure AI Speech for speech-to-text and text-to-speech; remember pipelines (speech then language).
  • Generative AI on Azure: Core Azure OpenAI concepts (prompts, completions/chat, embeddings at a concept level, grounding); responsible use practices (content filtering, data protection, human-in-the-loop, evaluation).

Exam Tip: If two services look plausible, ask: “Which one is marketed as the primary solution for this exact workload?” AI-900 favors first-party, purpose-built Azure AI services over generic tools.

Finish this sprint by re-reading your confusion list and rewriting each item as a clear “if you see X, choose Y” rule.

Section 6.6: Exam day checklist: identification, environment, stress control, and retake planning

Section 6.6: Exam day checklist: identification, environment, stress control, and retake planning

On exam day, you want zero preventable surprises. Your performance should be limited only by knowledge—not by logistics. Start with identification and check-in readiness. Confirm your legal name matches the registration, have acceptable ID ready, and complete system checks early if you are taking the exam online.

Prepare your environment for focus. For remote exams, clear your desk, ensure stable internet, and avoid second monitors or unnecessary peripherals that can trigger check-in issues. For test centers, arrive early and plan for storage rules (phones, watches, notes). Your goal is to reduce cognitive load before the timer starts.

Exam Tip: Use a deliberate “first 60 seconds” routine: read the first question slowly, identify the workload, and answer confidently. This prevents early panic and sets your pace.

Stress control is a skill. If you hit a confusing item, mark it, take one slow breath, and move on. AI-900 rewards breadth; you can regain points later. During review, avoid changing answers without a concrete reason (a new keyword you missed, a clearer service boundary, or a responsible AI principle mismatch). Random switching typically lowers scores.

Retake planning is part of professional exam strategy. If you do not pass, capture the domains that felt hardest immediately afterward, then rebuild with targeted practice: service boundaries, ML fundamentals, or generative AI responsibility. A structured retake plan turns a setback into a short, focused improvement cycle.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Final Review Sprint
  • Exam Day Checklist
Chapter quiz

1. A company wants to build a chatbot that answers employee questions using the company’s internal policy documents. They want the model to generate natural-language answers and cite relevant passages. Which Azure service best fits this requirement?

Show answer
Correct answer: Azure OpenAI Service
Azure OpenAI Service is used for generative AI scenarios (chat/completions) and can be combined with enterprise data to produce grounded answers. Azure AI Speech focuses on speech-to-text/text-to-speech and does not generate knowledge-based conversational responses. Azure AI Vision is for image analysis/OCR, not for a text-based generative chatbot.

2. You are reviewing a solution design: a model is trained once per week on labeled historical customer data. During the day, the website calls the model to return a prediction for each new customer. Which statement correctly describes training vs. inference?

Show answer
Correct answer: Training is when the model learns patterns from historical data; inference is when the trained model is used to make predictions on new data.
Training is the learning phase (fitting a model using historical data, often labeled). Inference is the scoring/prediction phase on new data. Option A reverses the definitions. Option C is incorrect because training and inference are distinct ML lifecycle stages regardless of where data is stored.

3. A retail team wants to predict next month’s sales revenue for each store using past sales and seasonal factors. Which machine learning task is this?

Show answer
Correct answer: Regression
Predicting a numeric value (revenue) is regression. Classification predicts a discrete category (e.g., 'high/low risk'). Clustering is unsupervised grouping without labeled targets and is not primarily used to predict a specific numeric outcome.

4. An insurance company needs to extract key-value pairs (e.g., policy number, customer name, effective date) from scanned application forms that follow a consistent layout. Which Azure AI service is the best fit?

Show answer
Correct answer: Azure AI Document Intelligence
Azure AI Document Intelligence is designed for document processing such as extracting structured data (forms, key-value pairs, tables) from documents. Azure AI Vision image classification identifies objects/categories in images, but it is not optimized for extracting structured fields from forms. Azure AI Speech handles audio workloads, not document field extraction.

5. A team is preparing for an AI-900 deployment review. They want to reduce the risk of the model producing unfair outcomes for different demographic groups and be able to explain model behavior to stakeholders. Which responsible AI principles most directly apply?

Show answer
Correct answer: Fairness and transparency
Fairness addresses unequal or biased outcomes across groups, and transparency/interpretability supports explaining model behavior. Reliability and performance focus on stability/accuracy under expected conditions but do not directly target bias or explainability. Privacy and security focus on protecting data and access, not primarily on fairness or interpretability.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.