HELP

+40 722 606 166

messenger@eduailast.com

AI-900 Practice Test Bootcamp: 300+ MCQs with Explanations

AI Certification Exam Prep — Beginner

AI-900 Practice Test Bootcamp: 300+ MCQs with Explanations

AI-900 Practice Test Bootcamp: 300+ MCQs with Explanations

Master AI-900 with focused domains, explanations, and full mock exams.

Beginner ai-900 · microsoft · azure · azure-ai-fundamentals

Prepare to pass Microsoft AI-900 with practice-first learning

This bootcamp is a practice-test driven course built for the Microsoft AI-900: Azure AI Fundamentals exam. If you’re new to certifications but have basic IT literacy, you’ll get a structured path through the official exam domains—paired with exam-style multiple-choice questions (MCQs) and clear explanations that teach you how to think like the test.

Instead of reading endless theory, you’ll train the skill that matters most on exam day: recognizing the workload, choosing the right Azure AI approach, and eliminating distractors. Each chapter includes targeted practice blocks designed to mirror the tone and decision-making of the real exam.

What the AI-900 exam covers (and how this course maps to it)

The course is structured as a 6-chapter book aligned to Microsoft’s published objectives:

  • Describe AI workloads — core AI concepts, scenarios, and responsible AI foundations.
  • Fundamental principles of ML on Azure — ML types, training lifecycle, evaluation, and Azure Machine Learning basics.
  • Computer vision workloads on Azure — common image analysis and OCR scenarios and service-selection cues.
  • NLP workloads on Azure — text analytics, language understanding, translation, and speech patterns.
  • Generative AI workloads on Azure — prompt concepts, grounding, safety, and Azure OpenAI fundamentals.

Course structure (6 chapters, built for beginners)

Chapter 1 gets you exam-ready operationally: registration and scheduling, how scoring works, what question formats to expect, and a study strategy that uses practice tests the right way (baseline → analyze → remediate → re-test). This is where you set up your plan and avoid common beginner mistakes.

Chapters 2–5 each focus on one or two official domains. You’ll learn the key terms and “selection logic” Microsoft tends to test (for example, identifying whether a scenario is classification vs regression, or recognizing when a vision/OCR solution is required). Every chapter includes domain practice sets with explanations so you can build confidence and speed.

Chapter 6 finishes with full mock exams, a weak-spot analysis workflow, and an exam-day checklist. You’ll leave with a clear idea of where you’re strong, where you’re guessing, and how to tighten up before your scheduled attempt.

Why this bootcamp helps you pass

  • Aligned to the official AI-900 domains—no filler.
  • Practice-first approach with explanations that teach decision-making, not memorization.
  • Beginner-friendly: assumes no prior certification experience and no coding required.
  • Mock exams plus final review to improve accuracy and pacing.

Get started on Edu AI

If you’re ready to train with domain-focused practice and track your progress, you can Register free to begin. You can also browse all courses to build a complete Microsoft fundamentals learning path.

What You Will Learn

  • Describe AI workloads and identify common Azure AI scenarios (Describe AI workloads)
  • Explain core machine learning concepts and how ML workloads are implemented on Azure (Fundamental principles of ML on Azure)
  • Choose Azure services for image and video analysis use cases (Computer vision workloads on Azure)
  • Choose Azure services for text analysis, conversational AI, and speech scenarios (NLP workloads on Azure)
  • Explain generative AI concepts and select Azure services and responsible AI practices (Generative AI workloads on Azure)

Requirements

  • Basic IT literacy (files, web apps, and cloud basics)
  • No prior certification experience required
  • No programming required (helpful but not necessary)
  • A computer with internet access to take quizzes and mock exams

Chapter 1: AI-900 Exam Orientation and Study Game Plan

  • Understand the AI-900 exam format and question types
  • Register for the exam and set up your test environment
  • Build a 2-week and 4-week study plan
  • How to use practice tests effectively (review loop + error log)

Chapter 2: Describe AI Workloads (Domain Deep Dive + MCQs)

  • AI workload types and when to use them
  • Responsible AI fundamentals for AI-900 scenarios
  • Azure AI service selection basics for common workloads
  • Domain practice set: 60+ AI workloads questions with explanations

Chapter 3: Fundamental Principles of Machine Learning on Azure (Domain + MCQs)

  • Core ML types, training lifecycle, and evaluation metrics
  • Supervised vs unsupervised vs reinforcement learning exam patterns
  • Azure ML concepts: datasets, compute, pipelines, and AutoML
  • Domain practice set: 80+ ML fundamentals questions with explanations

Chapter 4: Computer Vision Workloads on Azure (Domain + MCQs)

  • Image analysis capabilities and typical use cases
  • OCR and document processing basics for AI-900
  • Custom vision vs prebuilt vision: selection strategy
  • Domain practice set: 60+ computer vision questions with explanations

Chapter 5: NLP and Generative AI Workloads on Azure (Domains + MCQs)

  • Text analytics, translation, and language understanding essentials
  • Speech and conversational AI basics (bots and agents)
  • Generative AI foundations: prompts, grounding, and safety
  • Domain practice set: 90+ NLP + Generative AI questions with explanations

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist

Jordan Whitaker

Microsoft Certified Trainer (MCT)

Jordan Whitaker is a Microsoft Certified Trainer who helps beginners pass Microsoft fundamentals exams through domain-focused practice and clear explanations. He has designed exam-prep programs aligned to official Microsoft objectives, with an emphasis on test strategy and rapid remediation.

Chapter 1: AI-900 Exam Orientation and Study Game Plan

This bootcamp is built around one promise: you will not “study AI,” you will study what AI-900 tests. The AI-900: Microsoft Azure AI Fundamentals exam is a fundamentals credential, but it is still a certification exam—meaning it rewards precise service selection, correct terminology, and the ability to match a scenario to the right Azure AI approach. Your goal in Chapter 1 is to set expectations, remove logistics surprises, and lock in a repeatable study loop you’ll use through all 300+ practice questions.

AI-900 questions rarely ask you to implement code. Instead, they test whether you can recognize AI workload types (vision, language, conversational AI, anomaly detection, generative AI), understand core ML ideas (training vs inference, features/labels, classification vs regression), and pick the most appropriate Azure services (Azure AI Services, Azure AI Search, Azure OpenAI, Azure Machine Learning) while applying responsible AI principles. A strong candidate reads a scenario and immediately asks: “What is the workload? What is the data type? What service family fits? What responsible AI risk might be referenced?”

Use this chapter to set up your exam plan in two modes: a 2-week sprint and a 4-week steady plan. Both rely on the same engine: baseline test → error log → targeted review → re-test. The more disciplined your review loop, the less time you spend re-learning the same mistake.

Practice note for Understand the AI-900 exam format and question types: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set up your test environment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a 2-week and 4-week study plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How to use practice tests effectively (review loop + error log): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand the AI-900 exam format and question types: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set up your test environment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a 2-week and 4-week study plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How to use practice tests effectively (review loop + error log): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand the AI-900 exam format and question types: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: What AI-900 measures and how domains map to learning

AI-900 measures practical literacy across Azure AI workloads and the core concepts that support them. You are not expected to be a data scientist; you are expected to recognize which workload fits a business problem and which Azure service is designed for it. In other words: “know the why and which,” not the “build from scratch.”

Map your learning to the five outcome domains you will see repeatedly across practice tests: (1) Describe AI workloads (identify common scenarios such as recommendation, forecasting, anomaly detection, conversational bots); (2) Fundamental principles of ML on Azure (training vs inference, supervised vs unsupervised, evaluation metrics, overfitting); (3) Computer vision workloads on Azure (image classification, object detection, OCR, video analysis); (4) NLP workloads on Azure (sentiment, key phrases, NER, translation, speech-to-text); and (5) Generative AI workloads on Azure (prompts, grounding, embeddings, safety filters, responsible AI).

Exam Tip: Train yourself to translate the scenario into a “workload label” first. If the question says “extract text from invoices,” label it OCR; if it says “identify people and objects in a photo,” label it image analysis/object detection; if it says “classify emails,” label it text classification; if it says “generate a summary,” label it generative AI. Once labeled, Azure service choice becomes a short list, not a guess.

Common trap: confusing “Azure AI Services” (prebuilt APIs for vision/language/speech) with “Azure Machine Learning” (build/train/manage ML models). On AI-900, if the scenario is a standard capability (OCR, sentiment, translation) and no custom model training is required, the test usually wants Azure AI Services. If the scenario mentions creating a custom model, managing experiments, or tracking runs, Azure Machine Learning is a better fit.

As you study, keep a running “domain-to-service” map. This becomes your mental index during the exam and prevents time loss when multiple answers look plausible.

Section 1.2: Registration, scheduling, pricing, and ID requirements

Registering correctly is part of passing. Most candidate stress comes from avoidable logistics issues, not content. Schedule early enough to create a deadline, but not so early that you rush and under-practice. For this bootcamp, choose either a 2-week sprint date or a 4-week steady date, then plan backward.

Registration typically flows through Microsoft’s certification portal and an exam delivery provider. You will choose delivery method (online proctored vs test center), date/time, and confirm your legal name. Ensure the name on your registration matches your government-issued ID exactly (including middle name formatting, if present). A mismatch can block you from starting the exam.

Pricing can vary by region and discounts (student, employer programs, cloud skills challenges). Build a buffer for potential rescheduling fees and do not assume free retakes. If you plan to test online, confirm you have a quiet room, stable internet, and a supported device configuration. If you plan to test at a center, plan arrival time and parking/transportation.

Exam Tip: Decide your test mode based on your environment. If you cannot guarantee silence, uninterrupted time, and a clear desk at home, a test center is often less stressful even if it is less convenient.

  • Online proctoring checklist: clear desk, no extra monitors, webcam/mic working, allowed ID ready, phone put away unless instructed, stable network.
  • Test center checklist: arrive early, bring allowed ID, store personal items per center rules, know the check-in process.

Common trap: underestimating environment rules. Even innocent behaviors (reading aloud, looking away frequently, having notes visible) can trigger a proctor warning. Treat exam-day compliance as seriously as study prep.

Section 1.3: Scoring, passing, retakes, and exam-day policies

AI-900 is scored, timed, and policy-driven like any certification exam. The exam score report is typically scaled (not a simple percentage). That means you should avoid “target score myths” like “70% correct equals pass.” Your job is to consistently choose the best answer under exam conditions.

Passing requirements, retake rules, and waiting periods can change, so verify the current policy on the official certification page before scheduling. Build your plan assuming you want to pass on the first attempt; use retake knowledge as a safety net, not as a strategy.

Exam-day policies focus on security and fairness: no unauthorized materials, strict identity verification, and controlled breaks. For online proctoring, breaks may not be allowed or may be tightly regulated. For test centers, leaving the room may have restrictions. Plan hydration and comfort accordingly.

Exam Tip: Do a “policy rehearsal” the day before: confirm your ID, exam confirmation email, start time (including time zone), and the check-in window. Candidates miss exams simply by confusing time zones or arriving late to the online check-in.

Common trap: assuming you can review flagged questions indefinitely at the end. Some exam interfaces or question sets may limit backtracking in certain sections. Your time management must be resilient: answer the question in front of you as if you might not see it again, while still moving efficiently.

Also remember: AI-900 often uses scenario wording to test responsible AI awareness. If you see phrasing about fairness, privacy, transparency, or safety, do not treat it as “extra.” It can be the core of the scoring objective for that item.

Section 1.4: Question styles (MCQ, case, drag-drop) and time management

Expect a mix of multiple-choice questions (single best answer), multi-select items (choose all that apply), scenario/case-style questions, and matching/drag-and-drop formats. The content is fundamentals, but the question design often includes distractors that sound correct unless you notice one key constraint (for example: “needs custom training,” “must run on Azure,” “requires translation,” “needs OCR”).

Time management starts with recognizing your “decision point.” For most AI-900 items, you can decide in under 60–75 seconds if you use a consistent method: identify workload → identify data type → identify whether custom training is required → choose service. If you cannot decide quickly, mark it (if allowed), make your best selection, and move on. Do not sink minutes into one stubborn question and sacrifice multiple easy ones later.

Exam Tip: Read the last line first when it asks “Which service should you use?” Then scan the scenario for the constraint that eliminates options. Many distractors are eliminated by one word: “video,” “real-time,” “custom,” “structured documents,” or “generate.”

  • MCQ trap: two answers both “do AI,” but only one is the intended prebuilt API (Azure AI Services) versus a platform (Azure Machine Learning).
  • Multi-select trap: selecting a service because it’s popular rather than because it matches the workload. Treat each selection as “must be true,” not “could be useful.”
  • Drag-drop trap: mixing up training steps (data → training → evaluation → deployment/inference). Keep a simple pipeline in mind.

For case-style questions, avoid rereading the entire case repeatedly. Extract facts into a mini checklist: data type, goal, constraints, and what “success” looks like. Then answer each associated question against that checklist.

Section 1.5: Study strategy for beginners (notes, spaced repetition, labs optional)

If you are new to AI and Azure, your biggest risk is trying to learn everything. Your advantage is that AI-900 is built around patterns. Use a study strategy that favors repetition of those patterns rather than deep implementation detail.

Choose a 2-week plan if you can study 60–90 minutes per day with heavier practice-test volume. Choose a 4-week plan if you prefer 30–60 minutes per day with more review spacing. In both plans, split your time across (1) concept review, (2) service mapping, and (3) practice questions with explanations. Do not postpone practice tests until “after you finish learning”; the practice is how you learn what the exam actually emphasizes.

Exam Tip: Build a one-page “service cheat sheet” from memory every few days (not by copying). If you cannot recall which service fits OCR vs sentiment vs speech-to-text, that will show up immediately, and you can fix it early.

  • Notes: keep notes as decision rules (e.g., “prebuilt NLP → Azure AI Language,” “custom model lifecycle → Azure Machine Learning,” “RAG/search grounding → Azure AI Search”).
  • Spaced repetition: convert confusing terms into flashcards (features/labels, classification/regression/clustering, precision/recall, responsible AI principles).
  • Labs (optional): do small, guided experiences if you learn better by seeing the portal, but avoid turning labs into a time sink. AI-900 rewards recognition over configuration mastery.

Common trap: overfocusing on algorithms. You do need basic ML vocabulary, but you do not need to derive gradient descent. You are more likely to be tested on choosing the right metric or recognizing overfitting than on math.

Section 1.6: Practice-test workflow (baseline, analyze, remediate, re-test)

Your practice-test workflow will determine your score more than your total study hours. This course is a “300+ MCQs with explanations” bootcamp, so treat each question as a diagnostic tool. The objective is not to finish questions—it is to eliminate repeated mistakes.

Step 1: Baseline. Take an initial timed set to reveal your weak domains. Do not look up answers mid-test. Your baseline is only valid if it reflects exam conditions.

Step 2: Analyze. For every missed or guessed question, write an entry in an error log with four fields: (1) domain (AI workload, ML concepts, vision, NLP/speech, generative/responsible AI), (2) what you chose, (3) why it was tempting, (4) the rule you will use next time. This converts “I got it wrong” into “I changed my decision process.”

Exam Tip: Track “near-misses” (questions you got right but were unsure). Those are future wrong answers under pressure. Mark them and review the explanation anyway.

Step 3: Remediate. Study only what your error log proves you need: service boundaries, vocabulary, or scenario cues. Create a short checklist for that topic (for example, “OCR vs document understanding vs form extraction” or “classification vs regression”).

Step 4: Re-test. Re-test with new questions or reshuffled sets, timed. Your goal is to see the same pattern and apply the corrected rule automatically. Repeat the loop until your error log stops growing and starts stabilizing around a few rare edge cases.

Common trap: rereading notes instead of re-testing. Recognition without retrieval is not exam readiness. If you cannot answer under time pressure, you do not yet “own” the concept—no matter how clear it felt while reading.

Chapter milestones
  • Understand the AI-900 exam format and question types
  • Register for the exam and set up your test environment
  • Build a 2-week and 4-week study plan
  • How to use practice tests effectively (review loop + error log)
Chapter quiz

1. You are planning your AI-900 preparation. You want to align your study approach with what the exam actually measures rather than learning implementation details. Which activity best reflects the AI-900 exam focus?

Show answer
Correct answer: Practice mapping business scenarios to AI workload types and selecting the appropriate Azure AI service family
AI-900 is a fundamentals exam that emphasizes identifying AI workload types (vision, language, conversational AI, anomaly detection, generative AI), core ML concepts, responsible AI, and selecting the correct Azure AI services. Writing full implementations (B) and performance tuning/infrastructure optimization (C) are more aligned with associate-level engineering exams, not AI-900.

2. A company is taking a baseline practice test and notices recurring mistakes when choosing between Azure AI Services, Azure AI Search, Azure OpenAI, and Azure Machine Learning. Which study action is most likely to reduce repeat errors by using the recommended review loop?

Show answer
Correct answer: Maintain an error log that records the missed question, the correct service choice, and the reason the wrong option was tempting; then retest after targeted review
The chapter’s game plan is baseline test 92 error log 92 targeted review 92 re-test, specifically to stop re-learning the same mistake. Re-reading everything (B) is inefficient and not targeted. Skipping service-selection practice (C) conflicts with AI-900’s heavy emphasis on selecting the appropriate Azure AI approach for a scenario.

3. You are mentoring a learner who thinks AI-900 questions will require writing code to implement models. Which statement most accurately sets expectations for the AI-900 exam format and question types?

Show answer
Correct answer: Most questions are scenario-based and test recognition of workload type, core ML ideas, responsible AI concepts, and correct Azure service selection rather than coding
AI-900 typically assesses conceptual understanding and the ability to choose the right AI workload and Azure service for a scenario; it rarely asks you to implement code. Debugging scripts (B) and MLOps pipeline design (C) are beyond the fundamentals scope and are more typical of engineering-focused certifications.

4. A team is creating a repeatable method to answer AI-900 scenario questions during practice. Which first step best matches the recommended decision process in Chapter 1?

Show answer
Correct answer: Identify the workload and data type (e.g., vision vs language, text vs image) before selecting the Azure AI service family
A strong AI-900 approach is: determine the workload type and data type first, then map to the right service family and consider any responsible AI risks mentioned. Picking a product first (B) encourages confirmation bias and increases errors. Assuming one service is always best (C) is incorrect because AI-900 expects you to choose the most appropriate service (e.g., Azure AI Services vs Azure AI Search vs Azure OpenAI vs Azure Machine Learning) based on the scenario.

5. You are choosing between a 2-week sprint and a 4-week steady plan to prepare for AI-900. Regardless of the timeline, which sequence of activities is the core engine of the study plan described in Chapter 1?

Show answer
Correct answer: Baseline practice test 92 create/update an error log 92 targeted review of weak areas 92 re-test
Chapter 1 emphasizes a disciplined loop: baseline test, track mistakes in an error log, do targeted review, then re-test to confirm improvement. A single last-minute test (B) does not create feedback-driven improvement. Avoiding practice (C) undermines the scenario-recognition and service-selection skills AI-900 measures.

Chapter 2: Describe AI Workloads (Domain Deep Dive + MCQs)

This chapter targets the AI-900 objective area “Describe AI workloads and identify common Azure AI scenarios.” In practice tests, this domain is where candidates lose points not because the concepts are hard, but because the wording is subtle. You’ll be asked to classify a scenario (prediction vs perception vs language vs generation), pick the right Azure service category (prebuilt AI service vs custom ML), and recognize responsible AI concerns that the exam expects you to notice in everyday business prompts.

Use this chapter as a decision playbook: first identify the workload type, then map it to the Azure service family, then sanity-check the choice against responsible AI and common distractors. The sections that follow align to the lesson goals: workload types, responsible AI fundamentals, Azure AI selection basics, and a practice block (delivered separately as MCQs in Section 2.6) with a remediation checklist to close gaps quickly.

Exam Tip: When a prompt includes “predict,” “forecast,” “classify,” or “recommend,” think machine learning. When it includes “detect,” “recognize,” “extract,” “read,” “transcribe,” or “translate,” think prebuilt perception/language services. When it includes “create,” “generate,” “summarize,” “chat,” “compose,” think generative AI.

Practice note for AI workload types and when to use them: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Responsible AI fundamentals for AI-900 scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure AI service selection basics for common workloads: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 60+ AI workloads questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for AI workload types and when to use them: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Responsible AI fundamentals for AI-900 scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure AI service selection basics for common workloads: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 60+ AI workloads questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for AI workload types and when to use them: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Responsible AI fundamentals for AI-900 scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: AI concepts and workload categories (prediction, perception, language, generation)

Section 2.1: AI concepts and workload categories (prediction, perception, language, generation)

AI-900 tests whether you can quickly categorize a business problem into a workload type. Start by sorting the scenario into one of four buckets: prediction, perception, language, or generation. This first move eliminates most wrong answers because Azure services are organized around these workload families.

Prediction workloads produce a numeric value or class for a future/unknown outcome: churn risk, credit default, demand forecasting, fraud classification, or “which product is most likely to be bought.” These are typically machine learning problems and often require historical labeled data. On the exam, prediction is the “ML default” unless the prompt is clearly about analyzing unstructured media (images/audio/text) with prebuilt APIs.

Perception workloads interpret the world through sensors or media—images, video, or audio. Look for cues like “identify objects,” “detect faces,” “read text from images,” “analyze a video stream,” or “transcribe a call.” These are frequently addressed by Azure AI services (Vision, Speech) rather than building a custom model from scratch.

Language workloads focus on understanding or processing text: sentiment analysis, key phrase extraction, entity recognition, translation, question answering, summarization (extractive), and conversational bots. The exam often mixes “language understanding” with “conversation,” so watch for whether the goal is to extract insights from text (analysis) versus manage dialogue (bot) versus convert speech to text (speech).

Generation workloads create new content: drafting emails, generating code, producing images from prompts, or synthesizing a natural-language response in a chat interface. Generation is the hallmark of foundation models/LLMs and is a different pattern from traditional supervised learning.

Exam Tip: If the output is “new content” (a paragraph, an image, a rewritten document), that’s generation. If the output is “a label/score” (spam/not spam, 0–1 risk), that’s prediction. If the output is “recognized elements” (objects, faces, OCR text), that’s perception. If the output is “language insights” (sentiment, entities, translation), that’s language.

Section 2.2: Key AI terms (features, labels, inference, training, evaluation)

Section 2.2: Key AI terms (features, labels, inference, training, evaluation)

Expect vocabulary questions and “choose the best definition” items. The trap is that terms are related but not interchangeable. In AI-900, the core ML pipeline language shows up repeatedly, even in questions that look like they’re about Azure services.

Features are the input variables used by a model (e.g., customer tenure, purchase frequency, last login date). Labels are the target outputs you want the model to learn (e.g., churned: yes/no). A classic trap is reversing them. If the scenario says “predict if a patient will be readmitted,” then readmitted is the label; vitals and history fields are features.

Training is the process of fitting a model on data (often labeled). Inference is using the trained model to produce predictions on new data. On the exam, “real-time scoring,” “predict in production,” or “apply the model” means inference. “Build,” “fit,” “learn from,” or “create a model using historical data” means training.

Evaluation measures how well the model generalizes. You may see metrics implied rather than named (e.g., “minimize false negatives” in medical screening). Evaluation is where you detect overfitting (model performs great on training data but poorly on new data). Even if AI-900 doesn’t require deep metric math, it does expect you to know that you validate with held-out data and choose metrics appropriate to the problem.

Exam Tip: When a question mentions “ground truth,” it’s referring to labels or verified correct outcomes used for training/evaluation. If it mentions “features extracted from images/text,” that’s still features—just derived from unstructured inputs.

Also watch for classification vs regression language: classification predicts categories (approve/deny), regression predicts continuous values (price, temperature). In practice scenarios, “forecast” usually implies regression; “detect fraud” implies classification.

Section 2.3: Responsible AI principles and risk areas (fairness, reliability, privacy, transparency)

Section 2.3: Responsible AI principles and risk areas (fairness, reliability, privacy, transparency)

AI-900 includes responsible AI as a first-class objective, not an afterthought. The exam typically embeds responsible AI cues into otherwise normal scenarios (hiring, lending, healthcare, surveillance). Your job is to recognize the risk area and pick the mitigation or principle that matches.

Fairness concerns unequal model performance or outcomes across groups (e.g., higher loan denials for a protected class). Traps include assuming fairness is only about removing sensitive columns. Even if you remove “gender,” proxies (zip code, school) can reintroduce bias. The exam likes “evaluate performance across demographic groups” as the correct fairness-aligned action.

Reliability and safety focus on consistent performance, robustness, and handling edge cases. If a system must work under changing conditions (lighting changes in vision, accents in speech), reliability is the principle. Look for language like “fails intermittently,” “works in testing but not in production,” or “needs to handle unusual inputs.”

Privacy and security relate to protecting sensitive data and controlling access. Scenarios involving call recordings, medical notes, or customer PII are privacy-heavy. The exam often expects you to prefer least-privilege access, data minimization, and careful logging/retention policies rather than “store everything for later.”

Transparency is about explaining system behavior and making it clear when AI is used. If a prompt asks for “why did the model decide this?” or “users must understand limitations,” transparency is the right principle. In practice, transparency may include documentation, model cards, and communicating confidence/uncertainty.

Exam Tip: If the scenario touches people’s rights/opportunities (jobs, credit, healthcare), prioritize fairness and transparency. If it touches system correctness under stress (noise, drift, edge cases), prioritize reliability. If it touches sensitive data handling, prioritize privacy.

Section 2.4: When to use Azure AI services vs custom ML (rule of thumb selection)

Section 2.4: When to use Azure AI services vs custom ML (rule of thumb selection)

A frequent AI-900 question pattern: “Which approach/service should you use?” The exam is less about memorizing product names and more about selecting between prebuilt Azure AI services and custom machine learning (often implemented with Azure Machine Learning). Use a simple rule of thumb: choose prebuilt when the task is common and the output format is standard; choose custom ML when the task is domain-specific, requires proprietary labels, or needs tailored performance.

Use Azure AI services (prebuilt APIs) when you need OCR, image tagging, face detection, speech-to-text, translation, sentiment, entity extraction, or document layout extraction—especially when you don’t have training data or don’t want to manage the ML lifecycle. These services are designed for quick integration and are strong defaults on the exam.

Use custom ML when you must predict a business outcome unique to your organization (churn, risk, demand), when you have labeled historical data, or when off-the-shelf models won’t capture your domain. Also choose custom when you need to incorporate internal features (transactions, telemetry) that prebuilt services won’t know about.

Use generative AI services (Azure OpenAI) when the requirement is to generate, summarize, reason over, or chat with natural language—especially when “few-shot” prompting is implied. But don’t confuse generation with classic Q&A over a fixed FAQ: if the scenario is “answer from a curated knowledge base,” it may be a retrieval/Q&A pattern, not purely free-form generation.

Exam Tip: If the question says “without building or training a model,” that is a neon sign pointing to Azure AI services. If it says “train a model using historical data,” that points to custom ML. If it says “generate” or “compose,” that points to generative AI.

Section 2.5: Common exam scenarios and distractor patterns for AI workloads

Section 2.5: Common exam scenarios and distractor patterns for AI workloads

AI-900 distractors often look plausible because they belong to the right general area (AI) but the wrong workload type or wrong level of customization. Train yourself to spot the keyword that determines the workload.

Scenario: “Analyze images from a factory line to find defects.” The distractor is a prediction service (ML regression) when the better match is a vision/perception workload. If the prompt says “classify defect types from images,” it’s still perception, but you might need custom vision if defect categories are highly specific. If it says “read serial numbers from photos,” that’s OCR (perception) not ML prediction.

Scenario: “Transcribe customer calls and detect sentiment.” This is two workloads: speech recognition (audio→text) and language analysis (sentiment). The trap is picking only one. On multi-step scenarios, the exam expects you to map each step to the correct workload category.

Scenario: “Build a chatbot for HR policy questions.” Distractors include vision services or generic ML classification. Look for whether the bot must handle dialog (conversation) and whether answers come from a known corpus (retrieval/Q&A) versus open-ended generation. If “must cite company policy documents,” think retrieval grounded responses and responsible AI (reduce hallucinations).

Scenario: “Predict when a machine will fail.” This is prediction (often time-series/regression or classification). The distractor is anomaly detection versus forecasting: anomaly detection finds unusual behavior now; failure prediction estimates future risk/time-to-failure.

Exam Tip: When two answer choices are both “AI,” choose the one that requires the least bespoke training if the problem is standard (OCR, translation). Choose custom ML if the label is business-specific and derived from internal outcomes.

Section 2.6: Practice block A (exam-style MCQs) + remediation checklist

Section 2.6: Practice block A (exam-style MCQs) + remediation checklist

This chapter’s Practice Block A contains 60+ exam-style multiple-choice questions focused on workload identification, service-family selection, and responsible AI cues. Treat the practice set as a diagnostic: your goal is not just a score, but a clear map of which workload keywords and selection rules you miss under time pressure.

As you review explanations, label each missed question with one failure mode: (1) misclassified workload (prediction vs perception vs language vs generation), (2) confused training vs inference, (3) picked custom ML when a prebuilt service was implied (or vice versa), (4) missed a responsible AI risk cue, or (5) fell for a distractor that matched the input type but not the required output.

  • Remediation Checklist: Re-read the scenario and underline the output the business wants (label/score, extracted text, transcript, generated content).
  • Write the workload category in one word: prediction, perception, language, generation. If you can’t do it instantly, that’s the gap.
  • Identify whether the prompt implies “no model building” (prebuilt Azure AI services) or “train using historical data” (custom ML).
  • For people-impacting scenarios, add a responsible AI note: fairness, reliability, privacy, transparency. Pick one dominant risk.
  • Re-attempt the question without looking at choices; then match to the closest choice. This reduces distractor bias.

Exam Tip: On timed practice, don’t debate between two services by name. First lock the workload type and whether you need prebuilt vs custom. Names are secondary; the exam rewards correct mapping more than memorizing product catalogs.

After completing Practice Block A, aim for consistency: you should be able to explain why each wrong option is wrong in one sentence (wrong output, wrong input modality, requires training when the scenario forbids it, or ignores a responsible AI constraint). That one-sentence discipline is what translates into passing performance on test day.

Chapter milestones
  • AI workload types and when to use them
  • Responsible AI fundamentals for AI-900 scenarios
  • Azure AI service selection basics for common workloads
  • Domain practice set: 60+ AI workloads questions with explanations
Chapter quiz

1. A retail company wants to predict next week’s demand for each product based on historical sales, promotions, and seasonality. Which AI workload type best fits this scenario?

Show answer
Correct answer: Machine learning (forecasting/prediction)
This is a prediction/forecasting scenario (time-series style), which maps to a machine learning workload in the AI-900 domain. Computer vision is used for interpreting images/video (detect/recognize), and NLP translation is a language workload, neither of which addresses forecasting numeric demand.

2. A manufacturer wants to extract text from scanned PDF invoices and store the vendor name, invoice number, and total amount in a database. Which Azure AI service category is the best fit?

Show answer
Correct answer: Azure AI Document Intelligence (prebuilt document extraction)
Extracting and structuring data from documents (invoices, forms, receipts) aligns with Azure AI Document Intelligence. Image generation is a generative workload, not extraction. Training a custom ML model from scratch is typically unnecessary when a prebuilt document/OCR extraction service is designed for this exact workload and is the expected exam choice.

3. A support center wants a solution that can summarize long customer emails into a few bullet points for agents. Which workload type is being described?

Show answer
Correct answer: Generative AI (summarization)
Summarization is a classic generative AI language capability (create/compose/summarize). Computer vision focuses on analyzing images and video, not email text. Anomaly detection is a machine learning scenario for identifying unusual patterns in numeric/telemetry data, not producing summaries.

4. A bank deploys an AI model to recommend whether to approve or deny personal loans. Which Responsible AI consideration is most critical to evaluate for this scenario?

Show answer
Correct answer: Fairness (avoiding bias against protected groups)
Loan approval is a high-impact decision; the AI-900 Responsible AI fundamentals emphasize fairness and bias mitigation in such scenarios. Image artifact reliability is unrelated because the workload is not image generation/processing. Multi-language transcription inclusiveness is a different concern that applies to speech scenarios, not loan decisioning.

5. A company wants to automatically identify whether a product photo contains a damaged item (for example, a cracked screen) and flag it for review. Which Azure AI approach is most appropriate?

Show answer
Correct answer: Azure AI Vision (image analysis/custom vision-style classification/detection)
Detecting damage from photos is a computer vision workload, best aligned with Azure AI Vision capabilities for image classification/detection. Sentiment analysis applies to text opinions, not images. Speech-to-text applies to audio transcription, not product photos.

Chapter 3: Fundamental Principles of Machine Learning on Azure (Domain + MCQs)

This chapter targets the AI-900 skills measured area: Fundamental principles of machine learning (ML) on Azure. On the exam, you’re not expected to build deep models or write complex math. You are expected to (1) recognize ML problem types from short business scenarios, (2) understand the training lifecycle and why we split data, (3) pick the right evaluation metric and spot overfitting, and (4) map ML workloads to core Azure Machine Learning (Azure ML) concepts such as workspaces, compute, pipelines, and endpoints.

In practice tests, most missed questions come from mixing up problem types (classification vs regression; clustering vs anomaly detection) or confusing model training steps with deployment steps. This chapter also prepares you for the domain practice set (80+ questions) by showing you the patterns the exam uses to “hint” at the correct answer and the traps it uses to distract you.

Exam Tip: In AI-900, if the scenario emphasizes predicting a numeric value, it’s almost always regression; if it emphasizes choosing a category/label, it’s classification; if it emphasizes “grouping without labels,” it’s clustering; if it emphasizes “rare/unusual,” it’s anomaly detection. Read the verbs (“predict,” “classify,” “segment,” “detect unusual”) and the output type (number vs label vs group ID vs anomaly score).

Practice note for Core ML types, training lifecycle, and evaluation metrics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Supervised vs unsupervised vs reinforcement learning exam patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure ML concepts: datasets, compute, pipelines, and AutoML: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 80+ ML fundamentals questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Core ML types, training lifecycle, and evaluation metrics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Supervised vs unsupervised vs reinforcement learning exam patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Azure ML concepts: datasets, compute, pipelines, and AutoML: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 80+ ML fundamentals questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Core ML types, training lifecycle, and evaluation metrics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: ML problem types (classification, regression, clustering, anomaly detection)

AI-900 repeatedly tests whether you can match a business problem to a machine learning problem type. The exam often provides a short scenario (e.g., “predict,” “recommend,” “group,” “detect fraud”) and asks which ML approach fits. Focus on the expected output and whether labeled data exists.

Classification predicts a discrete label (yes/no, red/amber/green, product category). Binary classification is a common exam pattern (churn: leave vs stay; fraud: fraud vs not fraud). Multi-class classification appears in “identify the type of flower” or “route support tickets into categories.” A frequent trap is confusing classification with clustering: if categories are predefined and you have labeled examples, it’s classification; if categories must be discovered, it’s clustering.

Regression predicts a continuous numeric value (sales amount, temperature, time to failure, house price). The word “forecast” can be regression (numeric prediction) but sometimes the exam may frame forecasting as a general predictive task—still, if the output is a number, treat it as regression.

Clustering groups similar items when you do not have labels. Typical scenarios: customer segmentation, grouping news articles by topic, discovering device behavior patterns. The key cue is “no prior labels” or “find natural groupings.”

Anomaly detection identifies unusual events compared to normal patterns (fraud spikes, sensor outliers, unusual login behavior). Many solutions output an “anomaly score.” Trap: “defect detection” could be classification if you have labeled defects; it is anomaly detection if defects are rare and you mainly have normal examples.

Exam Tip: If the question mentions “training data with known outcomes,” you are in supervised learning (classification/regression). If it mentions “no labels” or “discover patterns,” it’s unsupervised (clustering). If it mentions an agent learning via rewards/penalties, it’s reinforcement learning—even if the answer choices try to lure you toward regression.

Section 3.2: Training lifecycle (data prep, split, train, validate, test, deploy)

The exam expects you to know the standard ML lifecycle and why each stage exists. A common AI-900 pattern is to ask which step prevents overfitting, which dataset is used for final evaluation, or what happens during deployment. Keep the sequence clear: prepare data → split → train → validate/tune → test → deploy → monitor.

Data preparation includes cleaning missing values, removing duplicates, encoding categories, normalizing/standardizing numeric features, and feature engineering. The exam won’t ask you for code, but it may test whether data quality affects model accuracy and fairness.

Data splitting is central: training data teaches the model; validation data supports hyperparameter tuning/model selection; test data is held out for final, unbiased performance estimation. A classic trap is using the test set to tune the model; that “leaks” information and inflates performance.

Training fits model parameters (weights, splits, coefficients). Validation compares candidate models and tunes hyperparameters (e.g., tree depth, regularization strength). Testing is the final check before deployment—if performance drops sharply on test data, suspect overfitting or dataset shift.

Deployment makes the model available for inference (real-time endpoint or batch scoring). Many candidates confuse “training” compute with “inference” compute; the exam may explicitly mention a “real-time endpoint” to indicate deployment.

Exam Tip: When you see “choose the best model” or “tune hyperparameters,” think validation set. When you see “estimate how the model will perform on new data,” think test set. When you see “make predictions from an application,” think deployed endpoint.

Section 3.3: Metrics and model evaluation (accuracy, precision/recall, MAE/RMSE, overfitting)

AI-900 focuses on selecting sensible metrics and recognizing evaluation pitfalls. You should know which metrics align with classification vs regression and when accuracy is misleading.

Accuracy (correct predictions / total) is common for classification but becomes a trap in imbalanced datasets (e.g., fraud detection). If only 1% of transactions are fraud, a model that predicts “not fraud” always is 99% accurate but useless.

Precision measures how many predicted positives are truly positive (reduce false positives). Recall measures how many actual positives are caught (reduce false negatives). The exam often frames this as a business trade-off: in medical screening, missing a disease (false negative) is costly → prioritize recall; in a high-friction fraud block that annoys customers → prioritize precision. You may also see F1 score as a balance between precision and recall.

For regression, know MAE (mean absolute error) and RMSE (root mean squared error). MAE is easier to interpret and is less sensitive to large outliers; RMSE penalizes large errors more heavily. If the scenario says “large errors are especially bad,” RMSE is often favored.

Overfitting appears when a model performs well on training data but poorly on new data. Exam cues include “high training accuracy, low test accuracy.” Typical mitigations: more data, simpler model, regularization, early stopping, cross-validation, and better feature selection. Another exam cue is data leakage—features that inadvertently encode the label (e.g., “refund issued” used to predict “fraud”).

Exam Tip: If the question emphasizes “rare positives,” suspect an accuracy trap and look for precision/recall. If it emphasizes “generalize to unseen data,” look for overfitting language and holdout testing/cross-validation as the fix.

Section 3.4: Azure Machine Learning overview (workspace, compute targets, endpoints)

On AI-900, you must recognize the core building blocks of Azure Machine Learning and how they map to training and deployment. The exam typically asks you to choose the right Azure ML component rather than implement it.

Azure ML workspace is the top-level resource where you organize experiments, models, datasets, compute, and deployments. Think of it as the “control plane” for ML assets and governance. If a question mentions “central place to manage models, runs, and artifacts,” the workspace is the answer.

Datasets/data assets represent curated data references used by training jobs and pipelines. The exam may use “versioning” or “reuse across experiments” as hints.

Compute targets provide the CPU/GPU resources for training or inference. Typical compute includes compute instances (interactive development), compute clusters (scalable training), and attached compute (e.g., AKS or other). Don’t overcomplicate: AI-900 mainly tests that training requires compute and may require GPUs for deep learning workloads.

Pipelines orchestrate repeatable ML workflows (prep → train → evaluate). In exam wording, pipelines show up as “automate and reproduce training steps” or “schedule retraining.”

Endpoints are how you deploy models for consumption. A real-time endpoint supports low-latency requests; batch scoring supports large volumes on a schedule. A trap is confusing “endpoint” with “training job”—endpoints serve predictions, not training.

Exam Tip: If the scenario is about “serving predictions to an app,” look for endpoint language. If it’s about “run training repeatedly with the same steps,” look for pipelines. If it’s about “manage everything in one place,” it’s the workspace.

Section 3.5: AutoML vs designer vs code-first—when each is appropriate

AI-900 expects you to distinguish three common ways to build ML in Azure ML: Automated ML (AutoML), the designer (drag-and-drop), and code-first development (SDK/CLI). Most exam questions are scenario-based: choose the approach that matches the team’s skills, the need for control, and the complexity of the task.

AutoML automatically tries algorithms and hyperparameters to find a strong baseline model for tasks like classification, regression, and time-series forecasting. It’s ideal when you want speed, a baseline, and minimal manual tuning. The trap: AutoML doesn’t remove the need for good data; poor features and leakage still break models. Also, AutoML is not a synonym for “no evaluation”—you still validate and test.

Designer provides a visual interface to build training pipelines with components. It’s good for learning, rapid prototyping, and teams that prefer a low-code approach. Exam cues: “drag-and-drop,” “visual workflow,” “no/low coding.” Trap: designer isn’t inherently more “enterprise”; it’s simply a development experience.

Code-first (Python SDK, MLflow-style tracking, scripts, notebooks) is best when you need full flexibility: custom data processing, custom models, advanced training loops, or integration into mature DevOps/MLOps workflows. Exam cues: “custom algorithm,” “full control,” “advanced tuning,” “integration with CI/CD.”

Exam Tip: When multiple answers seem plausible, pick based on constraints stated in the prompt: (1) need for rapid baseline → AutoML, (2) minimal coding/visual build → designer, (3) customization/control → code-first. If the prompt mentions “hyperparameter tuning across many models,” that’s a strong AutoML hint.

Section 3.6: Practice block B (exam-style MCQs) + wrong-answer analysis patterns

This chapter’s domain practice set (80+ ML fundamentals questions) will feel repetitive by design: AI-900 tests recognition of patterns more than deep implementation. To score consistently, treat each question as a classification problem of its own: identify the scenario type, map to the ML concept, then eliminate traps.

Wrong-answer pattern #1: Output-type mismatch. Distractors often swap classification and regression. If the output is a number (price, demand, duration), eliminate classification choices—even if the wording says “predict.” If the output is a label (approve/deny), eliminate regression metrics like MAE/RMSE.

Wrong-answer pattern #2: Supervised vs unsupervised confusion. If the prompt says “no labeled data,” remove classification/regression. If it says “historical data with known outcomes,” remove clustering as the primary approach. Many candidates miss the subtle phrase “known outcomes” or “ground truth.”

Wrong-answer pattern #3: Metric traps with imbalanced data. If positives are rare (fraud, defects, disease), accuracy is frequently the wrong choice. Look for precision/recall. If the business cost is “missing a positive,” choose recall; if it’s “too many false alarms,” choose precision.

Wrong-answer pattern #4: Lifecycle leakage. Options may imply tuning on the test set or evaluating only on training data. Prefer: train on training set, tune on validation, final check on test. If the question asks how to improve generalization, prefer cross-validation, regularization, or more data over “increase training accuracy.”

Wrong-answer pattern #5: Azure ML component confusion. Workspace vs compute vs endpoint vs pipeline appear together in options. Remember: workspace organizes, compute runs, pipeline automates steps, endpoint serves predictions.

Exam Tip: On timed practice, do a two-pass elimination: (1) remove answers that don’t match the output type or label availability, (2) remove answers that misuse lifecycle steps (test-set tuning, training-only evaluation), then pick the remaining option that best matches Azure terminology.

Chapter milestones
  • Core ML types, training lifecycle, and evaluation metrics
  • Supervised vs unsupervised vs reinforcement learning exam patterns
  • Azure ML concepts: datasets, compute, pipelines, and AutoML
  • Domain practice set: 80+ ML fundamentals questions with explanations
Chapter quiz

1. A retailer wants to predict the number of units of a product that will be sold next week for each store based on historical sales, promotions, and weather. Which ML task type should you use?

Show answer
Correct answer: Regression
This scenario requires predicting a numeric value (units sold), which is a regression problem. Classification is used to predict discrete labels (for example, 'high/medium/low demand'), not a continuous quantity. Clustering groups unlabeled data into segments and does not directly predict a numeric target.

2. You train a model in Azure Machine Learning and get 0.98 accuracy on the training set but only 0.62 accuracy on the test set. What does this most likely indicate?

Show answer
Correct answer: Overfitting
A large performance gap where training metrics are much better than test metrics is a common sign of overfitting (the model memorizes training patterns and fails to generalize). Underfitting typically produces poor results on both training and test data. Successful generalization would show similar performance on training and test data.

3. A bank wants to identify unusual credit card transactions in near real time. Fraud cases are rare compared to legitimate transactions. Which approach best matches this requirement?

Show answer
Correct answer: Anomaly detection
The key requirement is detecting rare/unusual events, which aligns with anomaly detection (often producing an anomaly score or flag). Clustering segments transactions into groups but is not specifically designed to surface rare events as the primary objective. Regression predicts a numeric value (for example, transaction amount) and does not directly address identifying unusual behavior.

4. You want to orchestrate a repeatable workflow in Azure Machine Learning that ingests data, trains a model, and evaluates it as a sequence of steps that can be rerun. Which Azure ML concept should you use?

Show answer
Correct answer: Pipeline
Azure ML pipelines are used to define and automate multi-step ML workflows (data prep, training, evaluation) so they are repeatable and traceable. A compute instance is a development compute resource used to run jobs/notebooks but does not define the workflow steps. A dataset is a way to reference and manage data, but it does not orchestrate end-to-end execution.

5. You are evaluating a binary classification model for detecting defective parts on an assembly line. Defective parts are uncommon, and missing a defect is very costly. Which metric should you prioritize?

Show answer
Correct answer: Recall
When false negatives are costly (missing actual defects), prioritize recall, which measures how many actual positives were correctly identified. MSE and R² are regression metrics for numeric prediction tasks and are not appropriate for evaluating a binary classification model.

Chapter 4: Computer Vision Workloads on Azure (Domain + MCQs)

Computer vision questions on AI-900 test whether you can recognize a vision problem statement and map it to the correct Azure service and capability. This chapter focuses on the “what” (workload types such as classification, detection, segmentation, and OCR), the “how” (what Azure AI Vision can do out of the box), and the “which” (how to select between prebuilt vision and custom models). You will also see how exam writers try to distract you with near-synonyms like “labeling,” “tagging,” “captioning,” “bounding boxes,” and “extract text,” each of which implies a different feature.

The exam rarely expects deep implementation detail (model architectures, training pipelines), but it does expect you to choose the right service family: Azure AI Vision for image analysis and OCR, Custom Vision for custom classification/detection, and to apply responsible AI thinking—especially around faces and identity. As you read, focus on cues in the scenario: is it about describing an image, finding objects, reading text, or recognizing a person? Those cues are your fastest path to the correct answer.

Exam Tip: Many incorrect options on AI-900 are “plausible” because multiple services can process images. Your job is to pick the most direct, purpose-built service for the asked outcome (for example, OCR → Azure AI Vision OCR/Read; custom product defect detection → Custom Vision).

  • Image analysis capabilities and typical use cases
  • OCR and document processing basics for AI-900
  • Custom vision vs prebuilt vision: selection strategy
  • Domain practice set: 60+ computer vision questions with explanations (in the practice block)

Use the sections below as an exam objective map: first identify the workload type, then align to Azure AI services, then eliminate traps using service selection logic and responsible AI constraints.

Practice note for Image analysis capabilities and typical use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for OCR and document processing basics for AI-900: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Custom vision vs prebuilt vision: selection strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 60+ computer vision questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Image analysis capabilities and typical use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for OCR and document processing basics for AI-900: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Custom vision vs prebuilt vision: selection strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 60+ computer vision questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Vision workload types (classification, detection, segmentation, OCR)

Section 4.1: Vision workload types (classification, detection, segmentation, OCR)

AI-900 vision questions begin with the workload type. If you can label the problem correctly, service selection becomes straightforward. The exam typically tests four core computer vision workloads: classification, object detection, segmentation, and OCR (optical character recognition).

Classification answers “What is in this image?” at the image level. Examples: “Is this a dog or a cat?” “Is this product packaging damaged: yes/no?” “Which brand logo is shown?” The output is a label (possibly with confidence). Classification is a match for prebuilt tagging/categorization (when generic) or Custom Vision classification (when domain-specific).

Object detection answers “Where are the objects?” It returns bounding boxes plus labels. Scenario cues: “count the people,” “locate cars,” “find defects on a part,” “detect a helmet on a worker.” Detection is often confused with classification; look for words like “locate,” “coordinates,” “bounding box,” or “in the image, identify each instance.”

Segmentation identifies pixel-level regions, not just boxes. On AI-900, segmentation appears less frequently, but you may see cues like “outline the object,” “mask,” “separate foreground from background,” or “precise boundaries.” In many fundamentals questions, segmentation is tested conceptually rather than through a specific Azure service choice.

OCR extracts text from images. This includes printed text, handwriting (depending on capability), and text embedded in photos (signs, labels). OCR is a different workload from “image description.” If the scenario mentions invoices, receipts, IDs, or scanned documents, OCR is usually the correct category.

Exam Tip: If the desired output includes any kind of “position” (box, region, coordinates), you’re no longer in pure classification. If the desired output is “text,” you’re in OCR—even if the input is a photo.

Common trap: “Tagging” is not OCR. Tagging labels objects/concepts; OCR returns characters/words/lines. If the question says “extract the serial number from a label,” tagging won’t help—OCR will.

Section 4.2: Azure AI Vision capabilities (image analysis, tagging, captions, object detection)

Section 4.2: Azure AI Vision capabilities (image analysis, tagging, captions, object detection)

Azure AI Vision (often referenced as “Azure AI Vision” or “Computer Vision” in exam materials) provides prebuilt capabilities for analyzing images without training a custom model. AI-900 expects you to recognize which prebuilt feature matches the scenario: image analysis, tagging, captions, and object detection are the most commonly tested.

Image analysis is a broad bucket: it can produce tags (keywords), a natural-language caption/description, identify certain visual attributes, and detect objects. The key exam skill is to match the requested output format to the feature name the question uses.

Tagging returns a list of labels describing the image contents (for example, “outdoor,” “person,” “bicycle”). Tagging is ideal when the user wants searchable metadata or simple categorization. If a scenario says “index images in a catalog so users can search,” tagging is a strong fit.

Captions produce a sentence-like description (for example, “A person riding a bike on a city street”). If the prompt mentions accessibility (screen readers), alt-text, or summarizing an image for users, captions are a classic answer.

Object detection identifies objects and returns bounding boxes. Use it when the question requires counts, localization, or highlighting items in an image feed (for example, drawing boxes around products on a shelf).

Exam Tip: When multiple options mention “Azure AI Vision,” read the verbs: “tag,” “caption,” “detect objects,” “extract text.” Those verbs map to distinct sub-features and are how the exam differentiates correct from partially correct answers.

  • Use tags for search/metadata and broad understanding.
  • Use captions for natural-language descriptions and accessibility.
  • Use object detection when location/boxes matter.

Common trap: A scenario may say “identify what’s in the picture” (which sounds like classification) but then ask to “highlight each item” (which implies detection). The requested output in the last sentence is usually the deciding clue.

Section 4.3: Optical character recognition and document scenarios (forms/receipts concepts)

Section 4.3: Optical character recognition and document scenarios (forms/receipts concepts)

OCR is a high-frequency AI-900 topic because it has clear, business-friendly scenarios: invoices, receipts, forms, IDs, labels, and scanned PDFs. The exam tests whether you can distinguish “read text” from “understand a document.” OCR extracts characters and words; document processing adds structure, fields, and meaning.

For fundamentals questions, expect phrasing like “extract text from images,” “read street signs,” “capture printed text from a PDF,” or “detect handwriting.” The baseline answer is typically Azure AI Vision OCR/Read capabilities. If the scenario mentions line-by-line text extraction or converting an image to text for downstream processing, that’s OCR.

When the scenario shifts to forms/receipts concepts, watch for cues like “total amount,” “vendor name,” “date,” “line items,” “key-value pairs,” or “table extraction.” Those cues move beyond raw OCR and into document understanding (often associated with Azure’s document processing services). Even if the question doesn’t name the product, the concept tested is: OCR gives you text; document intelligence gives you structured fields.

Exam Tip: If the requirement is “extract the receipt total,” the correct approach is not only OCR; you need a capability that understands layout/fields. If the requirement is “extract all text and store it,” OCR alone is sufficient.

  • OCR output: characters/words/lines (unstructured text).
  • Form/receipt output: structured fields (for example, Total, Tax, Merchant, Date) and sometimes tables.

Common trap: Questions may present an image analysis option (“tags/captions”) next to an OCR option. Tags/captions can describe that a “receipt” exists, but they will not reliably return the numeric values you need. Always align to the requested data type (text vs fields).

Also be careful with the input type. A scanned PDF is still “an image-like document” for OCR purposes. The exam won’t penalize you for not naming file formats; it will penalize you for picking a service that doesn’t extract text.

Section 4.4: Face and identity-related considerations (consent, privacy, responsible use)

Section 4.4: Face and identity-related considerations (consent, privacy, responsible use)

Face scenarios appear on AI-900 to test responsible AI understanding as much as technical mapping. The exam expects you to recognize that face-related analysis raises privacy, consent, and compliance requirements. You should be ready for questions that ask what you should do (or avoid) when working with face images.

Conceptually, there is a spectrum: detecting a face (presence and location), analyzing attributes (for example, whether a face is in the frame), and identifying or verifying a person (linking a face to an identity). The further you go toward identification, the stronger the ethical and regulatory constraints. On the exam, “identify a person” is a major red flag requiring explicit consent and careful policy alignment.

Exam Tip: If a scenario includes “employees,” “customers,” “students,” “public cameras,” or “retail shoppers,” immediately think: consent, transparency, data retention limits, and purpose limitation. These cues often indicate that the best answer mentions responsible use rather than a purely technical feature.

  • Consent: individuals should know and agree when biometric data is collected/used.
  • Privacy: minimize collection, secure storage, restrict access, define retention.
  • Fairness and harm: watch for biased outcomes; test and monitor.
  • Transparency: disclose when automated vision is used in decisions.

Common trap: Treating face recognition as “just another image classification.” Face and identity scenarios are frequently designed to see if you will overreach with identification when detection would satisfy the requirement (for example, “count people entering a room” needs detection/counting, not identity).

When in doubt, choose the least intrusive capability that satisfies the requirement, and pair it with governance controls. AI-900 rewards that conservative, responsible selection mindset.

Section 4.5: Service selection prompts and common exam traps for vision questions

Section 4.5: Service selection prompts and common exam traps for vision questions

This section is your “selection strategy” playbook: how to choose between prebuilt Azure AI Vision and Custom Vision, and how to avoid the most common distractors in vision MCQs. Most AI-900 service-selection questions can be solved by asking three prompts: (1) Do you need text, general image insights, or a domain-specific model? (2) Do you need localization (boxes/regions) or just a label/description? (3) Do you need to train with your own images?

Prebuilt Azure AI Vision is best when the scenario is generic and you want immediate analysis: tags, captions, object detection, and OCR. If the question implies “no training data,” “quickly add image analysis,” or “extract text,” prebuilt is usually the correct choice.

Custom Vision is best when the labels are specific to your business (custom product SKUs, proprietary defect types, brand-specific packaging states) and you can provide labeled images. It commonly appears in questions about “train a model to classify X” or “detect a custom object type not covered by prebuilt features.”

Exam Tip: The phrase “custom” in the scenario (custom parts, custom defects, proprietary categories) is a strong indicator for Custom Vision—especially if the question mentions providing training images or labeling.

  • If you must “train” or “label images,” think Custom Vision.
  • If you must “describe,” “tag,” or “caption,” think Azure AI Vision image analysis.
  • If you must “read text,” think Azure AI Vision OCR/Read (and document understanding if fields are needed).

Common trap: Confusing “object detection” with “image classification.” If the question wants to “locate each defect” or “draw boxes around items,” classification is incomplete. Another trap is picking a general ML service because it sounds powerful. AI-900 prefers the simplest purpose-built AI service over building your own model from scratch.

Common trap: Over-selecting a face/identity option. If the business goal is occupancy counting, queue length, or safety monitoring, you rarely need identity—so don’t choose identity-related answers unless the requirement explicitly says “verify/identify who the person is.”

Section 4.6: Practice block C (exam-style MCQs) + quick-reference decision tree

Section 4.6: Practice block C (exam-style MCQs) + quick-reference decision tree

This chapter’s domain practice set (Practice Block C) contains 60+ exam-style computer vision MCQs with explanations. Use them to drill the skill AI-900 actually measures: reading a short scenario, identifying the vision workload, then selecting the best Azure service/capability while eliminating distractors. Do not rush; your goal is to build a consistent, repeatable decision process.

Exam Tip: In practice, force yourself to underline (mentally) the required output type: “caption,” “tags,” “bounding boxes,” “text,” or “fields.” Then eliminate any option that doesn’t produce that output—even if it sounds related.

  • Step 1: Identify the output. Label? Boxes? Pixels? Text? Structured fields?
  • Step 2: Decide prebuilt vs custom. Generic understanding (prebuilt) vs domain-specific categories with training images (custom).
  • Step 3: Check responsible AI constraints. Faces/identity imply consent, privacy, and least-intrusive capability.
  • Step 4: Eliminate near-miss distractors. Tagging vs OCR, classification vs detection, OCR vs form extraction.

Quick-reference decision tree (memorize-level):

If the requirement is “extract text” → choose Azure AI Vision OCR/Read. If the requirement is “extract receipt totals/fields” → choose document processing/structured extraction rather than plain OCR. If the requirement is “describe an image in a sentence” → choose captions. If it is “add searchable keywords” → choose tagging. If it is “find and locate objects” → choose object detection. If it is “recognize custom categories/defects with your images” → choose Custom Vision.

Common trap: Some questions include multiple correct-sounding steps (for example, “use OCR and then classify the text”). Unless the scenario explicitly asks for a pipeline, AI-900 usually wants the single service capability that directly satisfies the requirement.

After completing Practice Block C, review your wrong answers and classify the mistake: workload misread (classification vs detection), output confusion (tags vs captions vs text), or service choice error (prebuilt vs custom). That error taxonomy is how you improve fastest before the exam.

Chapter milestones
  • Image analysis capabilities and typical use cases
  • OCR and document processing basics for AI-900
  • Custom vision vs prebuilt vision: selection strategy
  • Domain practice set: 60+ computer vision questions with explanations
Chapter quiz

1. A company is building a mobile app that must return a natural-language description of a user’s photo (for example, “a person riding a bicycle on a city street”). The company wants to use a prebuilt capability and avoid training a model. Which Azure service/capability should they use?

Show answer
Correct answer: Azure AI Vision Image Analysis (caption generation)
Azure AI Vision Image Analysis provides prebuilt features such as captions and tags for describing an image. Custom Vision classification requires training and is intended for custom categories, not general captioning. Face (identify) is for face-related scenarios (and has responsible AI restrictions); it does not generate general image descriptions.

2. A finance team needs to automatically extract printed and handwritten text from scanned invoices and return the text content for downstream processing. Which capability should you recommend?

Show answer
Correct answer: Azure AI Vision OCR/Read
OCR/Read is the purpose-built capability for extracting text from images and documents. Custom Vision object detection locates objects with bounding boxes and is not designed to extract text. Image tags describe visual content (for example, “invoice,” “paper”) but do not return the recognized text.

3. A manufacturer wants to detect whether a specific type of defect is present on its products from photos taken on an assembly line. The defect types are unique to the manufacturer’s products, and the company can label example images. Which approach is the best fit?

Show answer
Correct answer: Train a Custom Vision model (custom classification or detection, depending on whether location is needed)
When the categories are domain-specific (custom defects) and you have labeled images, Custom Vision is the appropriate choice for custom classification or object detection. Prebuilt Image Analysis captions/tags are generic and often won’t reliably detect specialized defects. OCR/Read is for extracting text, not identifying visual defects.

4. A retail company wants to build a solution that draws bounding boxes around every product visible on a shelf image and returns the coordinates for each product. Which workload type does this describe?

Show answer
Correct answer: Object detection
Bounding boxes and coordinates indicate object detection. Image classification assigns labels to the entire image (for example, “shelf”) without locations. OCR extracts text characters/words and is unrelated to locating non-text objects.

5. A company asks for a system that can identify a specific person by comparing their face against a database of known employees. You must recommend a service while considering responsible AI expectations for AI-900. What is the most appropriate guidance?

Show answer
Correct answer: Do not recommend face identification as a default; consider whether the scenario is allowed and use Face only if requirements and access policies permit
AI-900 expects you to recognize that face identification is a sensitive capability with responsible AI constraints and may require special approval/policies; you should not treat it as a routine recommendation. Image Analysis tags do not identify individuals by name. Custom Vision classification is not the appropriate guidance for face identification and does not remove the need to address responsible AI, privacy, and policy requirements.

Chapter 5: NLP and Generative AI Workloads on Azure (Domains + MCQs)

This chapter maps directly to the AI-900 skills measured around NLP workloads and generative AI workloads on Azure. On the exam, you are rarely asked to design deep architectures; instead, you are expected to recognize workload types (text analytics vs. translation vs. Q&A vs. speech), select the correct Azure service family (Azure AI Language, Speech, Bot Service, Azure OpenAI), and apply responsible AI basics (safety, privacy, and grounding).

You should practice spotting “keyword cues” in questions. For example: “extract key phrases,” “detect sentiment,” “recognize entities,” “summarize,” and “translate” almost always indicate an Azure AI Language capability. “Transcribe call center audio” indicates Speech to text. “Generate an email draft” or “chat with company policy grounding” indicates Azure OpenAI with grounding (retrieval) and safety controls.

Exam Tip: When two answers both sound plausible, choose the one that is the most managed and purpose-built for the described task. AI-900 favors Azure’s prebuilt AI services over custom ML unless the scenario explicitly requires training a custom model.

As you move into the domain practice set for NLP + Generative AI, keep a consistent approach: (1) identify the workload (classification, extraction, conversation, generation), (2) match to the Azure service, and (3) verify constraints like real-time vs batch, multilingual needs, and responsible AI requirements.

Practice note for Text analytics, translation, and language understanding essentials: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Speech and conversational AI basics (bots and agents): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Generative AI foundations: prompts, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 90+ NLP + Generative AI questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Text analytics, translation, and language understanding essentials: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Speech and conversational AI basics (bots and agents): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Generative AI foundations: prompts, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Domain practice set: 90+ NLP + Generative AI questions with explanations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Text analytics, translation, and language understanding essentials: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: NLP workload types (sentiment, key phrases, entities, summarization, translation)

Section 5.1: NLP workload types (sentiment, key phrases, entities, summarization, translation)

AI-900 expects you to recognize common NLP tasks from short scenario descriptions and map them to the right capability. Typical “text analytics” workloads include sentiment analysis (positive/negative/neutral), key phrase extraction (important terms in a document), entity recognition (people, places, organizations, dates, medical terms), and summarization (condense long text into shorter text). Translation is a separate—but often adjacent—need, and exam questions may mix translation with downstream analytics.

Look for phrasing cues: “analyze reviews to determine customer satisfaction” implies sentiment; “identify main topics in support tickets” implies key phrases; “extract order numbers and customer names” implies entity extraction; “create an executive summary from meeting notes” implies summarization; “convert English product descriptions to French” implies translation.

  • Sentiment analysis: Often used for social media, reviews, and surveys.
  • Key phrase extraction: Tagging or search indexing, topic surfacing.
  • Entity recognition: Compliance, redaction prep, routing, enrichment.
  • Summarization: Long documents, call transcripts, case notes.
  • Translation: Multilingual apps, localization, cross-lingual support.

Exam Tip: If a question says “translate speech,” don’t default to text translation. Think end-to-end: Speech-to-text (Speech service) → translate text → optionally text-to-speech.

Common trap: Confusing “language understanding” (intent classification) with “text analytics” (sentiment/entities). If the scenario is about predicting what a user wants (intent), you’re in conversational/language understanding territory, not sentiment.

Section 5.2: Azure AI Language and typical scenarios (classification, extraction, Q&A patterns)

Section 5.2: Azure AI Language and typical scenarios (classification, extraction, Q&A patterns)

Azure AI Language is the umbrella service family that covers many text-focused capabilities. For AI-900, the key is recognizing which scenario fits: classification (assign a label), extraction (pull structured info), and Q&A (return best answers from a knowledge base). In exam items, classification might appear as “route incoming emails into billing/technical/sales,” while extraction appears as “find names, locations, and dates from documents.”

Q&A patterns often show up as: “Create a customer support FAQ bot that answers from internal documentation.” The exam wants you to think of a managed Q&A approach rather than training a large custom model from scratch. Watch for phrasing like “knowledge base,” “FAQ,” “documents,” “support articles,” and “best answer.”

  • Classification: Intent routing, ticket categorization, moderation labels.
  • Extraction: Entities, key phrases, PII detection/redaction preparation.
  • Q&A: Retrieve the most relevant answer from curated content.

Exam Tip: If the scenario emphasizes “structured answers from existing content,” choose a Q&A/knowledge base solution. If it emphasizes “generate a new answer in a human tone,” that shifts toward generative AI (Azure OpenAI) possibly with retrieval grounding.

Common trap: Picking “Custom ML” for a standard categorization problem. Unless the question states you must build/train a custom model with features, AI-900 typically rewards choosing Azure AI Language’s built-in or configurable options.

Section 5.3: Speech workloads (speech-to-text, text-to-speech) and use-case mapping

Section 5.3: Speech workloads (speech-to-text, text-to-speech) and use-case mapping

Speech scenarios are heavily pattern-based on AI-900. If you see “transcribe,” “caption,” “convert audio to text,” or “call center recordings,” you should think speech-to-text. If you see “read aloud,” “voice output,” “natural sounding voice,” or “IVR voice,” think text-to-speech. Many real solutions chain both directions, but exam questions usually highlight one primary requirement.

Speech-to-text is commonly used for live captions, meeting notes, searchable call transcripts, and accessibility. Text-to-speech is used for voice assistants, audio versions of articles, and speaking responses from bots. Some questions add constraints like “multiple languages” or “real-time transcription,” which should still map to Speech capabilities rather than text analytics alone.

  • Speech-to-text cue words: dictation, captions, transcription, subtitles, call recording.
  • Text-to-speech cue words: voice, synthesize, speak, audio output, narration.

Exam Tip: If the question includes audio input, your first step is usually Speech. Don’t pick Azure AI Language just because “text analysis” is mentioned—Speech is what converts the modality from audio to text.

Common trap: Assuming text-to-speech is “translation.” It’s not. Translation changes language; text-to-speech changes output modality. You may combine both, but they solve different problems.

Section 5.4: Conversational AI basics (bots/virtual agents) and routing user intents

Section 5.4: Conversational AI basics (bots/virtual agents) and routing user intents

Conversational AI on AI-900 is less about coding bots and more about identifying the building blocks: a channel-facing bot/agent, an NLP layer to detect intent, and an orchestration/routing step to trigger the right action (answer a question, create a ticket, check an order). Scenarios commonly describe “a chatbot on a website,” “a virtual agent in Teams,” or “automate customer support.”

Intent routing is a repeated exam theme. You may be asked to distinguish between: (1) extracting information from a message (entities like order number), (2) classifying what the user wants (intent like “track order”), and (3) responding (static Q&A vs dynamic generation). The best answers usually separate these steps mentally: intent detection → entity extraction → fulfillment.

  • Bot/agent layer: Handles conversation flow, channels, and state.
  • Intent recognition: Routes to billing vs technical vs sales.
  • Fulfillment: Calls APIs, queries knowledge bases, or escalates to humans.

Exam Tip: If a scenario says “handoff to a human agent,” look for options that support escalation and conversation context retention. Many wrong answers ignore the operational requirement and focus only on text analysis.

Common trap: Treating a bot as the intelligence itself. A bot framework/service is often the container; language understanding, Q&A, speech, or generative models provide the intelligence.

Section 5.5: Generative AI concepts and Azure OpenAI basics (prompting, tokens, grounding)

Section 5.5: Generative AI concepts and Azure OpenAI basics (prompting, tokens, grounding)

Generative AI questions on AI-900 focus on core concepts and when to choose Azure OpenAI. Generative models create new content (text, code, summaries, explanations) based on prompts. On the exam, prompts are often described as “instructions,” “system message,” or “few-shot examples.” You should know that models operate on tokens (chunks of text), and that longer prompts and longer outputs consume more tokens, affecting latency and cost.

Grounding is a high-value concept: it means anchoring model responses in trusted data (company policies, product docs, ticket history) to reduce hallucinations and improve relevance. Many scenarios describe “use our internal documents” or “answer based only on company policy.” That is a grounding cue: retrieve relevant content and include it as context for the model rather than relying purely on the model’s general knowledge.

  • Prompting: Clear instructions, constraints, examples, and desired format.
  • Tokens: Input + output budget; impacts cost/limits.
  • Grounding: Retrieve-then-generate using enterprise data to improve accuracy.

Exam Tip: If the requirement is “generate” (draft, rewrite, brainstorm, chat), pick Azure OpenAI. If the requirement is “extract” (entities, key phrases), pick Azure AI Language. Don’t overuse generative AI for deterministic extraction tasks unless the scenario explicitly asks for freeform generation.

Common trap: Assuming generative AI automatically knows your company’s private data. It doesn’t—unless you provide the data via grounding or fine-tuning (where applicable). AI-900 commonly tests that you must supply context to get domain-specific answers.

Section 5.6: Responsible generative AI (safety, content filtering, privacy) + practice block D

Section 5.6: Responsible generative AI (safety, content filtering, privacy) + practice block D

Responsible AI is not optional on AI-900, and generative AI makes it more visible. Expect questions about safety controls, content filtering, and privacy boundaries. Safety focuses on preventing harmful outputs (hate, violence, sexual content, self-harm), reducing jailbreak success, and monitoring for policy violations. Privacy focuses on protecting sensitive inputs/outputs, controlling data access, and ensuring you do not leak personal or confidential information.

In practice, you apply responsible generative AI through layered controls: good prompting (clear constraints like “answer only from provided sources”), grounding (reduce hallucinations), access control to data sources, logging/monitoring, and content filtering/moderation policies. Exam scenarios may mention “users can type anything,” “public-facing chat,” or “employees entering customer data”—these are signals to prioritize filtering and privacy.

  • Safety: Use content filtering/moderation to reduce harmful output.
  • Privacy: Limit who can access prompts, responses, and grounded data.
  • Reliability: Ground responses and require citations when appropriate.

Exam Tip: If the scenario highlights compliance (PII, regulated data) the best answer usually includes privacy controls and data governance, not just “use a stronger model.” Stronger models do not replace policy.

Common trap: Treating “responsible AI” as only an ethical statement. On AI-900, it’s operational: content filtering, access control, and grounded responses are practical mechanisms.

Now move to Practice Block D (NLP + Generative AI domain set). Use a consistent elimination strategy: identify whether the scenario is analytics (extract/classify), conversation (intent + routing), speech (audio in/out), or generation (create new content). Then check for responsible AI constraints—public exposure, sensitive data, and required reliability—before you lock in the service choice.

Chapter milestones
  • Text analytics, translation, and language understanding essentials
  • Speech and conversational AI basics (bots and agents)
  • Generative AI foundations: prompts, grounding, and safety
  • Domain practice set: 90+ NLP + Generative AI questions with explanations
Chapter quiz

1. A retail company wants to automatically detect sentiment and extract key phrases from customer reviews in multiple languages. They want a managed service with prebuilt models and minimal custom training. Which Azure service should they use?

Show answer
Correct answer: Azure AI Language
Azure AI Language provides prebuilt NLP capabilities such as sentiment analysis and key phrase extraction (and supports multilingual scenarios). Azure Machine Learning is used to build/train custom models and is unnecessary when prebuilt text analytics meets the requirement. Azure Bot Service is for building conversational experiences and does not directly provide sentiment/key phrase extraction as the core workload.

2. A call center needs to convert recorded customer calls into searchable text transcripts to support compliance audits. The solution must handle audio input and return text. Which Azure capability best fits?

Show answer
Correct answer: Speech to text in Azure AI Speech
Speech to text in Azure AI Speech is purpose-built to transcribe audio into text. Azure AI Language text analytics services operate on text, not raw audio recordings. Document Intelligence is designed to extract structured data from documents (such as PDFs/images/forms), not transcribe spoken audio.

3. A company wants a website chatbot that answers employee questions using an internal HR policy document set. The chatbot must generate natural language answers grounded in those documents. Which approach aligns best with Azure-managed services for this generative AI workload?

Show answer
Correct answer: Use Azure OpenAI with retrieval/grounding to company documents
Azure OpenAI is the Azure service family for generative AI text generation, and grounding (retrieval over internal documents) is the recommended approach to answer with company policy context. Azure AI Speech focuses on speech workloads (transcription/translation/synthesis) rather than grounding a chat experience in documents. Training an LLM from scratch in Azure Machine Learning is not the AI-900-typical or most managed approach and is unnecessary for a Q&A chatbot scenario.

4. An application must translate product descriptions from English to Japanese in near real time using a prebuilt API. Which Azure service capability is the best match?

Show answer
Correct answer: Translation capability in Azure AI Language
Azure AI Language includes translation capabilities as a managed, purpose-built service for language translation. While Azure OpenAI can be prompted to translate, AI-900 guidance favors purpose-built managed services when the task is explicitly translation. Azure Bot Service provides bot orchestration and conversation logic, but it is not itself the translation engine.

5. A financial services firm is deploying a generative AI assistant for employees. They want to reduce the risk of harmful or policy-violating content in responses. Which control is most appropriate to apply in this scenario?

Show answer
Correct answer: Use content filtering/safety features for generative AI responses
Content filtering/safety features help detect and block harmful or policy-violating outputs, aligning with responsible AI expectations for generative AI workloads. Disabling grounding typically increases hallucination risk and reduces reliability, which works against safe, accurate responses. Storing prompts/completions in plain text can increase privacy and compliance risk; responsible AI emphasizes appropriate data handling rather than indiscriminate logging.

Chapter 6: Full Mock Exam and Final Review

This chapter is your transition from “learning the objectives” to “executing under exam conditions.” AI-900 rewards breadth, not deep coding ability. Your job is to recognize which Azure AI capability matches a scenario, explain core concepts in plain language, and avoid common traps where two answers look plausible. We’ll use two full-length mock exams (Part 1 and Part 2), then convert your results into a weak-spot plan, and finish with an exam day checklist you can actually follow.

The exam objectives you’ve practiced across this course—AI workloads, machine learning fundamentals on Azure, computer vision services, NLP/speech/conversational services, and generative AI plus responsible AI—show up as short scenario prompts with product selection or concept identification. The most consistent differentiator between pass and fail is pacing and disciplined elimination. In this chapter, you’ll practice those habits deliberately.

Exam Tip: Treat mock exams as diagnostics, not validation. A “good” mock score is one where you learn exactly why you missed questions and which keywords fooled you. The fastest improvement comes from fixing misunderstanding patterns (e.g., mixing up Azure Machine Learning vs Azure AI services, or Vision Studio vs Document Intelligence).

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam instructions, timing, and score targets

Section 6.1: Mock exam instructions, timing, and score targets

Your goal is to simulate the real AI-900 experience closely enough that exam day feels familiar. Use a quiet setting, one sitting, and no notes. If you must pause, restart the section from the beginning later—your brain needs the stamina training. Two full mock exams in this chapter are designed to cover mixed domains, mirroring how the certification blends AI workloads, ML fundamentals, vision, NLP/speech, and generative AI.

Timing: allocate a fixed window and follow it strictly. Plan a first pass where you answer everything you can quickly, mark uncertain items mentally, and return only if time remains. Your pacing target should be “steady, not perfect”: don’t overinvest time in a single ambiguous prompt when other items are likely straightforward.

Score targets: aim for a consistent practice score that gives you buffer for exam-day variance. If you’re below target, don’t repeat mock exams immediately—do weak-spot remediation first, then retest. Use a simple tracking sheet by domain (AI workloads, ML, vision, NLP, generative AI/responsible AI) and record (1) what you picked, (2) what the correct concept/service was, and (3) which keyword should have triggered it.

Exam Tip: In AI-900, many wrong answers are “real Azure products” but not the best fit. Your mock strategy should include identifying the primary task (classify vs extract vs generate vs search) before you even look at options.

Section 6.2: Full-length mock exam set 1 (mixed domains, exam difficulty)

Section 6.2: Full-length mock exam set 1 (mixed domains, exam difficulty)

Mock Exam Part 1 is built to match typical AI-900 difficulty: short scenarios, lightweight terminology, and a heavy emphasis on selecting the right Azure service. As you take this set, keep your focus on “what is the workload?” before “what is the product?” For example, if the scenario describes extracting structured fields from invoices or forms, your mental category should be “document understanding,” which points you toward Azure AI Document Intelligence rather than generic OCR.

Expect domain mixing: a prompt might mention chat, then include a requirement like “transcribe calls,” which quietly shifts part of the solution to Speech. The exam tests whether you can separate components: conversational interface (Azure AI Bot Service), language understanding (Azure AI Language), speech-to-text or text-to-speech (Azure AI Speech), and orchestration/integration (often implied, not asked).

Also watch for ML fundamentals slipping into service selection. If the scenario says “train a model using your labeled data and manage experiments,” that is an Azure Machine Learning signal. If it says “use a prebuilt model to detect objects in images,” that’s Azure AI Vision. The trap is assuming everything is “Azure Machine Learning” because it sounds advanced. AI-900 expects you to recognize when you are consuming a prebuilt AI capability versus building/training a custom model.

Exam Tip: For Part 1, practice elimination: remove options that are (a) non-AI data services, (b) developer tools not asked for, or (c) AI services that solve a different modality (vision vs language vs speech). If two answers remain, choose the one that directly meets the scenario’s core verb: “extract,” “classify,” “recognize,” “translate,” or “generate.”

Section 6.3: Full-length mock exam set 2 (mixed domains, new question bank)

Section 6.3: Full-length mock exam set 2 (mixed domains, new question bank)

Mock Exam Part 2 uses a new question bank to prevent memorization and to expose gaps that Part 1 might not reveal. Here, you should deliberately apply your “keyword-to-service” map under time pressure. When you see “summarize,” “draft,” “rewrite,” or “ground responses in company data,” immediately classify it as generative AI. The follow-up decision is often: do you need a general LLM capability (Azure OpenAI Service) and do you need retrieval/grounding patterns (commonly implemented via Azure AI Search with your content)? The exam may not demand architecture detail, but it will test whether you pick the correct family of services.

Part 2 also tends to surface responsible AI and model evaluation concepts. If a scenario references fairness, transparency, privacy, reliability, or preventing harmful outputs, your answer should align with responsible AI practices (e.g., content filtering, human review, data minimization, monitoring) rather than purely technical performance metrics. A frequent trap is selecting “increase training data” for every issue; that helps accuracy but does not automatically address bias or safety.

For classic ML prompts, anchor on fundamental terms: features, labels, training vs inference, classification vs regression vs clustering. If the question asks about predicting a numeric value, that’s regression; predicting a category is classification; grouping unlabeled data is clustering. On AI-900, getting the concept right is often enough to eliminate two distractors.

Exam Tip: When you finish Part 2, don’t just tally a score. Tag each miss as either “concept confusion” (didn’t know) or “execution error” (misread, rushed, swapped similar services). Your remediation approach differs: concept confusion needs re-study; execution error needs strategy (slower reading, highlight key nouns/verbs mentally).

Section 6.4: Final review by domain (AI workloads, ML, vision, NLP, generative AI)

Section 6.4: Final review by domain (AI workloads, ML, vision, NLP, generative AI)

Now convert your mock results into a final domain review. Start with AI workloads: be able to distinguish common patterns—anomaly detection, forecasting, personalization/recommendations, computer vision, NLP, and conversational AI. The exam often tests whether you can identify the workload type from the business phrasing. If the prompt says “detect unusual spikes,” think anomaly detection; if it says “predict next month’s demand,” think forecasting (a regression/time-series style use case).

Machine learning fundamentals on Azure: know the lifecycle words (ingest → prepare → train → evaluate → deploy → infer/monitor) and basic metric intent (accuracy for classification, RMSE/MAE for regression). Also know when Azure Machine Learning is the right answer: custom model training, experiment tracking, pipelines, and model management. Trap: confusing AutoML (a feature for automating model selection/training) with “no training required” AI services.

Computer vision: separate image classification, object detection, OCR, and document extraction. Azure AI Vision covers many prebuilt vision tasks; OCR is not the same as form field extraction, which is where Document Intelligence is typically the best fit. Trap: selecting Vision for invoices when the scenario emphasizes structured fields like totals, dates, vendor names.

NLP and speech: identify tasks such as sentiment analysis, key phrase extraction, named entity recognition, language detection, translation, and speech-to-text/text-to-speech. Azure AI Language maps to most text analytics and language understanding tasks; Azure AI Speech maps to audio and voice. Trap: treating “chatbot” as a single service—often it’s Bot Service plus Language/Speech depending on modality.

Generative AI and responsible AI: recognize LLM-driven generation, summarization, question answering, and prompt-based workflows, usually tied to Azure OpenAI Service. Know responsible AI practices: mitigate bias, protect privacy, ensure transparency, and add safeguards like content filtering and human-in-the-loop for sensitive workflows. Trap: confusing “search” with “generate”—retrieval (Azure AI Search) finds documents; generation (Azure OpenAI) creates new text. Many real solutions combine both, but the exam usually asks for the primary capability.

Section 6.5: Top mistakes and last-minute mnemonic/decision guides

Section 6.5: Top mistakes and last-minute mnemonic/decision guides

Your last-minute goal is to prevent avoidable misses. The most common mistakes on AI-900 are (1) picking a real but suboptimal service, (2) mixing up similar terms (OCR vs document extraction; classification vs clustering), and (3) ignoring a single constraint word like “custom,” “real-time,” “prebuilt,” or “structured fields.” Train yourself to spot constraint words first.

Use quick decision guides. For ML type: “Number = Regression, Name = Classification, No labels = Clustering.” For service selection: “Prebuilt capability = Azure AI service; Train/manage your own model = Azure Machine Learning.” For modality: “Image/video = Vision; audio = Speech; text understanding = Language; structured documents = Document Intelligence; generation = Azure OpenAI.” These aren’t perfect in every edge case, but they win most exam scenarios quickly.

Another trap is overthinking architecture. AI-900 is fundamentals: it rarely expects multi-service pipelines unless the scenario explicitly demands multiple modalities (e.g., voice + translation + bot). If the prompt is simple, your answer should be simple. If you’re torn between two options, ask: which one directly states the capability in its name or standard use case?

Exam Tip: When two answers feel right, choose the one that minimizes “extra work.” For example, if you need invoice field extraction, Document Intelligence typically requires less custom training and parsing than a generic OCR output you must post-process heavily.

Section 6.6: Exam day checklist (environment, pacing, elimination strategy)

Section 6.6: Exam day checklist (environment, pacing, elimination strategy)

On exam day, your objective is consistent execution. Prepare your environment: stable internet, a quiet room, and a cleared desk. If you’re testing remotely, follow proctor rules strictly—unexpected interruptions can cost time and focus. If you’re testing at a center, arrive early and plan for check-in delays.

Pacing plan: do a fast first pass—answer what you know, and don’t “wrestle” with a question. Your second pass is where you re-read the prompt carefully and apply elimination. The exam often gives you enough information, but it’s easy to miss one keyword that flips the correct service (e.g., “extract key-value pairs” vs “read printed text”).

Elimination strategy: first eliminate wrong modality (vision vs language vs speech). Next eliminate wrong level (prebuilt AI service vs custom ML training). Finally, match to the scenario verb: analyze, detect, extract, translate, transcribe, or generate. This three-step approach reduces “50/50” guesses into high-confidence picks.

Mindset: don’t chase perfection. If you hit a confusing item, mark your best choice and move on. Spending three minutes to gain one point is rarely worth it if it causes you to rush five later questions.

Exam Tip: Before submitting, do a brief sanity scan for classic swaps: classification vs regression, Vision OCR vs Document Intelligence extraction, Language vs Speech, and Azure Machine Learning vs Azure AI services. These are the errors that most often occur under time pressure.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
Chapter quiz

1. A company wants to add AI to an app that identifies objects in photos and returns labels (for example, “bicycle,” “dog,” “tree”). The team does not want to build or train a custom model. Which Azure service should they use?

Show answer
Correct answer: Azure AI Vision
Azure AI Vision provides prebuilt computer vision capabilities such as image tagging and object detection without requiring custom training. Azure Machine Learning is used to build/train/manage custom ML models and is more work than needed for a prebuilt labeling scenario. Azure AI Document Intelligence focuses on extracting structured data from documents (forms, invoices, receipts), not general object labeling in images.

2. You are reviewing a teammate’s plan for an AI-900 exam. They say: “I’ll focus on deep coding details and SDK methods because that’s what the exam tests.” Which response best aligns with the AI-900 exam focus?

Show answer
Correct answer: AI-900 emphasizes recognizing Azure AI workloads and selecting the correct service for a scenario, not writing code.
AI-900 is a fundamentals exam that rewards breadth: understanding common AI workloads and matching scenarios to Azure AI capabilities. Deep coding implementation details and hyperparameter optimization are not the primary focus (those align more with practitioner/engineer roles and other certifications). GPU/Kubernetes scaling is also beyond AI-900’s core objectives.

3. A retail company wants to build a solution that answers customer questions using a chat interface and can route complex issues to a human agent. They prefer a managed service designed for conversational experiences. Which Azure option best fits?

Show answer
Correct answer: Azure AI Bot Service
Azure AI Bot Service (and related conversational capabilities) is designed to build and connect chatbots and integrate with channels, with logic for escalation/hand-off patterns. Azure AI Vision is for image/video analysis, not chat. Azure AI Speech is for speech-to-text/text-to-speech and speech translation; it can complement a bot but is not the primary service for building the conversational workflow.

4. During mock exams, a learner repeatedly confuses Azure Machine Learning with Azure AI services (for example, selecting Azure Machine Learning for prebuilt OCR). Which weak-spot plan is most effective?

Show answer
Correct answer: Create a keyword-to-service mapping (prebuilt API capabilities vs custom model lifecycle) and drill mixed scenario questions focused on service selection.
A targeted weak-spot plan for AI-900 should address misunderstanding patterns: Azure AI services provide prebuilt models via APIs, while Azure Machine Learning is for building, training, and managing custom ML models and pipelines. Memorizing SDK classes is unnecessary for AI-900. Ignoring the weakness is risky because service-selection traps are common across domains and can cause repeated misses.

5. On exam day, you encounter several questions where two answers seem plausible. What strategy best matches recommended exam technique for AI-900-style scenario prompts?

Show answer
Correct answer: Identify scenario keywords, eliminate clearly incorrect services first, then choose the remaining option that best matches the workload.
AI-900 questions often hinge on keywords that map to a workload/service; disciplined elimination is effective when options look similar. The “longest answer” heuristic is unreliable and not exam guidance. Marking everything without selecting initially risks time management and can lead to unanswered questions; a better approach is to answer when possible and flag only genuinely uncertain items.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.