HELP

+40 722 606 166

messenger@eduailast.com

AI-900 Mock Exam Marathon: Timed Sims & Weak Spot Repair

AI Certification Exam Prep — Beginner

AI-900 Mock Exam Marathon: Timed Sims & Weak Spot Repair

AI-900 Mock Exam Marathon: Timed Sims & Weak Spot Repair

Timed AI-900 simulations plus targeted drills to fix what costs points.

Beginner ai-900 · microsoft · azure-ai-fundamentals · exam-prep

AI-900 pass prep built around timed simulations

AI-900 (Microsoft Azure AI Fundamentals) rewards broad understanding, fast scenario recognition, and the ability to pick the right Azure AI approach under time pressure. This course is designed as a “mock exam marathon”: you’ll practice in timed conditions, diagnose what’s costing points, and then repair weak areas with focused drills aligned to the official exam domains.

You’ll start with an exam-ready foundation: how to register, what to expect in Microsoft’s exam interface, how scoring works, and how to build a realistic study plan even if you’ve never taken a certification exam before. From there, Chapters 2–5 map directly to the AI-900 domains and train you to answer the exam’s most common prompt patterns.

Mapped to official AI-900 exam domains

This blueprint follows the domains exactly as listed in the official objectives:

  • Describe AI workloads
  • Fundamental principles of machine learning on Azure
  • Computer vision workloads on Azure
  • NLP workloads on Azure
  • Generative AI workloads on Azure

Each domain chapter combines concept clarity (what Microsoft expects you to know) with exam-style practice focused on scenario selection, terminology, and common distractors.

How the 6-chapter structure helps you improve fast

Chapter 1 sets your strategy: exam registration and logistics, study scheduling, time management, and a baseline diagnostic so you know where to focus. Chapters 2–5 go domain by domain, using “service selection” thinking (what workload is being described, what output is required, and what Azure capability best fits). You’ll also practice responsible AI considerations where they naturally appear in questions—especially around safety, privacy, transparency, and appropriate use of generative AI.

Chapter 6 is the capstone: two timed mock exam parts, followed by a structured review workflow to turn mistakes into repeatable rules. You’ll finish with a short exam-day checklist to manage pacing, reduce second-guessing, and avoid avoidable errors.

Best fit for beginners

This course assumes basic IT literacy but no prior certification experience and no coding requirement. If you’re new to AI, you’ll learn the core vocabulary (training vs inference, features vs labels, common metrics) and how to translate a business scenario into an Azure AI approach.

Get started on Edu AI

If you’re ready to begin, Register free and start your first timed sprint. Prefer to compare options first? You can also browse all courses on the platform.

By the end, you won’t just “know” AI-900 topics—you’ll be practiced at answering Microsoft-style questions quickly, reviewing errors efficiently, and walking into exam day with a clear plan.

What You Will Learn

  • Describe AI workloads: identify common AI solution types and responsible AI considerations
  • Fundamental principles of ML on Azure: choose ML approaches, understand training vs inference, and interpret core metrics
  • Computer vision workloads on Azure: select Azure services for image analysis, detection, OCR, and vision scenarios
  • NLP workloads on Azure: choose Azure services for text analytics, language understanding, translation, and conversational AI
  • Generative AI workloads on Azure: explain LLM concepts, prompt patterns, and Azure OpenAI use cases and safety basics

Requirements

  • Basic IT literacy (web apps, cloud concepts, and using a browser)
  • No prior certification experience required
  • No coding required (helpful but optional to recognize basic ML terms)
  • A computer with reliable internet access to take timed simulations

Chapter 1: AI-900 Game Plan—Registration, Scoring, and Strategy

  • Understand AI-900 format, question types, and time management
  • Register for the Microsoft AI-900 exam and set a test date
  • Build a 2-week and 4-week study plan mapped to official domains
  • Baseline diagnostic: identify strengths and weak spots

Chapter 2: Describe AI Workloads—Core Concepts and Scenario Selection

  • Recognize AI workload types and when to use each
  • Apply responsible AI principles to exam scenarios
  • Practice set: workload selection and governance questions
  • Timed mini-sim: Describe AI workloads domain sprint

Chapter 3: Fundamental Principles of ML on Azure—Train, Evaluate, Deploy

  • Choose the right ML approach: supervised, unsupervised, and reinforcement learning
  • Interpret core ML concepts: features, labels, overfitting, and metrics
  • Select Azure tooling for ML workflows (conceptual, exam-level)
  • Practice set: ML fundamentals and Azure ML scenario questions

Chapter 4: Computer Vision Workloads on Azure—Pick the Right Service

  • Identify core computer vision tasks and outputs
  • Map scenarios to Azure vision services and capabilities
  • Practice set: OCR, image analysis, and detection questions
  • Timed mini-sim: computer vision domain sprint

Chapter 5: NLP + Generative AI Workloads on Azure—Language to LLMs

  • Choose Azure NLP services for analytics, translation, and conversations
  • Understand generative AI basics: LLMs, prompts, grounding, and safety
  • Practice set: NLP and generative AI mixed questions
  • Timed mini-sim: language and generative AI domain sprint

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist

Nadia Herrera

Microsoft Certified Trainer (MCT)

Nadia Herrera is a Microsoft Certified Trainer who builds beginner-friendly certification paths focused on real exam objectives and time-efficient practice. She has coached hundreds of learners through Microsoft Fundamentals exams using mock exams, targeted remediation, and exam-day strategies.

Chapter 1: AI-900 Game Plan—Registration, Scoring, and Strategy

AI-900 (Azure AI Fundamentals) rewards clarity over complexity. Your job in this course is not to become a data scientist—it’s to become predictable under exam conditions: understand what Microsoft is testing, recognize common distractors, and manage time across mixed question formats. This “Mock Exam Marathon” approach starts with logistics (registration and scheduling), then locks in a scoring-aware strategy, and finally builds a repeatable practice loop that exposes weak spots quickly.

Across the exam, you’ll be asked to describe AI workloads, distinguish training vs. inference, select Azure services for computer vision and NLP, and explain generative AI and Azure OpenAI safety basics. You don’t need deep math; you do need crisp definitions, appropriate service selection, and responsible AI thinking that threads through every domain. This chapter sets your plan: book the exam date, choose a two- or four-week path mapped to official domains, and run a baseline diagnostic so you know where to spend your minutes.

Exam Tip: Treat AI-900 like a “terminology + scenario matching” test. When you miss items, it’s often because you chose a plausible-sounding service or metric, not because you lacked advanced knowledge.

Practice note for Understand AI-900 format, question types, and time management: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the Microsoft AI-900 exam and set a test date: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a 2-week and 4-week study plan mapped to official domains: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Baseline diagnostic: identify strengths and weak spots: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand AI-900 format, question types, and time management: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the Microsoft AI-900 exam and set a test date: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a 2-week and 4-week study plan mapped to official domains: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Baseline diagnostic: identify strengths and weak spots: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand AI-900 format, question types, and time management: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the Microsoft AI-900 exam and set a test date: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: Exam overview—Azure AI Fundamentals scope and audience

AI-900 is designed for broad audiences: business stakeholders, students, and technical practitioners who need a foundation in AI concepts and Azure AI services. The exam is not a coding test. Instead, it measures whether you can describe common AI solution types (classification, regression, clustering, anomaly detection), understand the basic machine learning lifecycle (training vs. inference), and select the right Azure service for common workloads in computer vision, natural language processing, and generative AI.

Expect scenario-style prompts that ask “what should you use?” rather than “how do you implement it?” For example, you may be given a need to extract text from scanned receipts (OCR), detect objects in images, analyze sentiment in customer reviews, or build a chatbot. The exam tests that you know which service family fits: Azure AI Vision for image analysis/OCR, Azure AI Language for text analytics, and Azure OpenAI for large language model use cases.

Common trap: confusing similarly named offerings or overcomplicating the solution. If the scenario is straightforward (e.g., “extract printed text”), do not jump to custom model training. Fundamentals questions usually favor managed, prebuilt capabilities unless the scenario explicitly requires custom labeling, bespoke categories, or model training.

Exam Tip: When a question mentions “training data,” “labels,” or “model versioning,” it’s pointing to custom ML concepts. When it mentions “quickly analyze,” “prebuilt,” or “no ML expertise,” it’s pointing to prebuilt Azure AI services.

Section 1.2: Registration steps—Microsoft Learn/Cert portal, scheduling, and policies

Your study plan becomes real when you register and pick a date. Schedule first, then study toward the clock. Register through Microsoft Learn and the Microsoft Certification dashboard (Cert portal), where you can locate AI-900 and launch scheduling with the exam provider. You’ll choose delivery (online proctored or test center), language, time zone, and an appointment time.

Plan for the policies that can derail test day: ID requirements, check-in timing, room/desk rules for online proctoring, and reschedule/cancellation windows. Online exams typically require a clean desk, stable internet, and a private room—no second monitor, notes, or phone access. If you cannot guarantee a compliant environment, a test center may reduce stress and eliminate last-minute technical risk.

  • Pick a date within your chosen 2-week or 4-week plan window.
  • Confirm name on your profile matches your ID exactly.
  • Read reschedule rules and set a calendar reminder before the deadline.
  • For online proctoring, run the system test and plan a wired connection if possible.

Exam Tip: Schedule the exam at your best cognitive time (often late morning). You’re training for timed performance, not just knowledge—timing matters.

Common trap: waiting to schedule “until you feel ready.” In exam prep, readiness follows commitment. A fixed date improves focus and prevents the endless “one more resource” loop.

Section 1.3: Scoring, passing, and how Microsoft exams are structured

Microsoft certification exams use scaled scoring. The headline number you care about is the passing score (commonly 700 on a 1–1000 scale). Do not assume each question is worth the same. Some items may be weighted differently, and some exams include unscored items used for validation. The practical takeaway: focus on consistent correctness across domains rather than trying to “game” point values.

Microsoft exams typically mix question types: multiple choice, multiple response (“choose all that apply”), drag-and-drop matching, and case-based scenario sets. You must manage attention switching. Multi-select items are a frequent score leak because candidates apply single-answer habits and miss that partial correctness may not be awarded—treat them as all-or-nothing unless the exam interface states otherwise.

Time management is part of the format. You need a pace that leaves room for review without rushing the last block. Build a simple rule: on first pass, answer what you can quickly, flag the rest, and return with remaining time. However, don’t over-flag—flagging is for uncertainty, not perfectionism.

Exam Tip: Read the last line first (“What should you use?”) before reading the scenario details. This prevents you from getting lost in irrelevant context and helps you identify what the question is truly testing: service selection, ML concept, metric interpretation, or responsible AI principle.

Common trap: overreading “Azure” branding and choosing a tool because it sounds advanced. Fundamentals questions reward the simplest correct service aligned to the requirement.

Section 1.4: Study strategy—domain weighting mindset and active recall

AI-900 preparation should be mapped to the official skill domains (AI workloads, machine learning fundamentals, computer vision, NLP, and generative AI). Your study time should mirror likely exam emphasis: you want coverage across all domains, with extra focus where you’re weakest. This course’s “Weak Spot Repair” approach starts with a baseline diagnostic (Section 1.5), then reallocates effort using your error log.

Choose a plan length based on your starting point:

  • 2-week plan: For candidates with some Azure familiarity or prior AI exposure. Emphasize daily timed practice plus targeted review.
  • 4-week plan: For true beginners. Spend the first half building vocabulary and service recognition, then increase timed simulation volume.

Use active recall, not passive rereading. Build a one-page “service map” you can recite: which Azure service handles OCR, object detection, sentiment, key phrase extraction, translation, conversational agents, and generative text. Also include core ML terms: features vs. labels, training vs. inference, supervised vs. unsupervised, and common metrics like accuracy, precision, recall, and F1.

Exam Tip: When you study metrics, always pair them with the risk they manage. Precision reduces false positives; recall reduces false negatives. The exam often frames this as a business consequence (e.g., missing fraud vs. flagging legitimate transactions).

Common trap: spending too long on model algorithms. AI-900 rarely demands algorithm detail; it demands scenario-fit reasoning and correct terminology.

Section 1.5: Timed practice approach—review loops and error logs

This course is a “Mock Exam Marathon,” which means your primary learning engine is timed simulation followed by disciplined review. Start with a baseline diagnostic: one timed mini-exam under realistic conditions. The goal is not a high score—it’s a clean signal about strengths and weak spots across the domains.

After each timed set, run a two-loop review:

  • Loop 1 (Immediate): Re-check every flagged question and every incorrect answer. Identify the exact failure mode: concept gap, service confusion, misread requirement, or time pressure.
  • Loop 2 (Repair): Create an error-log entry with (a) what the question was testing, (b) why the correct option fits, (c) why your choice was wrong, and (d) the “trigger phrase” you will watch for next time (e.g., “extract printed text” → OCR in Vision).

Time management practice must be intentional. Use a pacing target (questions per minute) and practice making “good enough” decisions. If two options look close, anchor on the requirement: is it prebuilt analysis, custom training, conversational interaction, or generative creation? The exam rewards alignment to requirement more than breadth of features.

Exam Tip: Your error log is your score multiplier. Most candidates repeat the same confusion—Vision vs. Language, LUIS-era terms vs. newer Language capabilities, or training vs. inference. Fix the pattern once and you gain points repeatedly.

Common trap: reviewing only wrong answers. Review lucky guesses too; they are hidden weaknesses that often flip under time pressure.

Section 1.6: Responsible AI mindset—how it appears across domains

Responsible AI is not a standalone topic you can “cram at the end.” In AI-900 it appears across domains as principles and practical safeguards. You should be able to recognize fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability—and identify what actions support them in real solutions.

In machine learning scenarios, responsible AI often shows up as questions about bias in training data, explainability, and monitoring. If the dataset underrepresents a group, fairness risk increases; if stakeholders need to understand why a prediction happened (loan approval, healthcare), transparency matters. In computer vision, consider privacy and consent when analyzing faces or personal imagery. In NLP, watch for toxicity, sensitive data leakage, and overreliance on sentiment as “truth.”

Generative AI and Azure OpenAI add safety basics: content filters, grounded responses, prompt design to reduce hallucinations, and careful handling of personally identifiable information. You should be able to articulate why you would apply moderation, limit data exposure, and log/monitor outputs.

Exam Tip: If a question asks “what should you do first” in a responsible AI context, the best answer is often about defining requirements and risks, assessing data quality, or adding human oversight—not choosing a technical feature.

Common trap: treating responsible AI as optional governance. Microsoft exams frame it as a core design requirement. When two options both “work,” choose the one that reduces harm, improves transparency, or strengthens privacy/security while meeting the scenario goal.

Chapter milestones
  • Understand AI-900 format, question types, and time management
  • Register for the Microsoft AI-900 exam and set a test date
  • Build a 2-week and 4-week study plan mapped to official domains
  • Baseline diagnostic: identify strengths and weak spots
Chapter quiz

1. You are planning your AI-900 preparation. You want a strategy that best matches how the exam is typically assessed. Which approach should you prioritize?

Show answer
Correct answer: Focus on crisp definitions and mapping scenarios to the correct Azure AI service and workload type
AI-900 is an Azure AI Fundamentals exam focused on identifying AI workloads, distinguishing concepts like training vs. inference, and selecting appropriate Azure AI services. Option A aligns with the exam’s domain expectations (terminology + scenario matching). Option B is incorrect because deep mathematical derivations are not a primary requirement at the fundamentals level. Option C is incorrect because AI-900 does not require building models from scratch or advanced coding skills; it emphasizes selecting managed services and understanding concepts.

2. A company sets a strict time budget for the AI-900 exam because it includes mixed question formats. Which time-management action best supports certification-style success?

Show answer
Correct answer: Do a baseline diagnostic first, then allocate study/practice time toward weak domains and rehearse pacing across mixed question types
AI-900 performance improves when you identify weak spots early and practice under exam-like pacing, since the exam tests recognition and selection under time constraints. Option A matches the recommended loop: baseline diagnostic, targeted remediation, and timed practice. Option B is wrong because delaying practice reduces familiarity with question patterns and common distractors. Option C is wrong because ignoring timing removes an important exam skill: managing minutes across question formats.

3. During a practice session, you notice you often confuse training and inference in scenario questions. Which scenario best represents inference rather than training?

Show answer
Correct answer: Using a pre-trained model to classify new images uploaded by users in real time
Inference is the act of using an already trained model to generate predictions on new data, which is described in option A. Option B describes training because it involves learning/adjusting parameters from labeled data. Option C is part of data preparation and dataset creation, which supports training but is not inference.

4. You are creating a 2-week or 4-week study plan mapped to the official AI-900 domains. Which set of topics best reflects what you should expect across the exam according to the chapter summary?

Show answer
Correct answer: AI workloads, training vs. inference, Azure services for computer vision and NLP, and generative AI plus Azure OpenAI safety basics
Option A matches the official-scope fundamentals described: workloads, training vs. inference, service selection for vision/NLP, and responsible/safe generative AI concepts (including Azure OpenAI safety basics). Option B is incorrect because AI-900 does not emphasize advanced math or low-level neural network implementation. Option C is incorrect because deep MLOps/Kubernetes engineering is beyond fundamentals and is not central to AI-900 domain coverage.

5. After a baseline diagnostic, you find you missed several questions because multiple answers looked plausible (for example, selecting the wrong Azure AI service). What is the most appropriate next step to reduce this specific error pattern?

Show answer
Correct answer: Create a review sheet that maps common scenarios to the correct workload/service and explicitly notes typical distractors and why they are wrong
AI-900 frequently tests scenario-to-service matching and uses plausible distractors. Option A directly addresses the root cause by reinforcing distinctions and building recognition of common traps, aligning with the exam’s terminology-and-scenario style. Option B is wrong because removing practice reduces exposure to distractor patterns and does not build exam readiness. Option C is wrong because it creates domain gaps; the exam spans multiple domains (workloads, training vs. inference, vision/NLP services, and responsible generative AI concepts).

Chapter 2: Describe AI Workloads—Core Concepts and Scenario Selection

In AI-900, “Describe AI workloads” is less about building models and more about selecting the right solution type for a business scenario. The exam repeatedly tests whether you can translate a short prompt—often written in non-technical language—into the correct AI workload category, the typical approach (rules vs ML vs deep learning), and the appropriate Azure service family (Azure Machine Learning, Azure AI Vision, Azure AI Language, Azure AI Search, Azure OpenAI, and bot services). This chapter builds your scenario-selection reflexes: recognize the workload, apply responsible AI principles where the prompt hints at risk, and avoid distractors that sound sophisticated but don’t match the task.

You’ll see the same pattern across items: identify inputs (text, images, audio, tabular data), outputs (label, score, bounding box, summary, answer), and constraints (latency, explainability, privacy, human-in-the-loop). Then choose the simplest workload that satisfies the requirement. Exam Tip: When two answers both “could work,” AI-900 usually rewards the most direct managed service aligned to the scenario (for example, OCR → Vision rather than a custom ML pipeline), unless the prompt explicitly demands custom training or domain-specific data.

This chapter also sets up your timed mini-sim mindset: you’ll practice quick workload recognition and governance cues (responsible AI) under time pressure, then use the “weak spot repair” loop—review why a distractor was tempting and what keyword should have redirected you.

Practice note for Recognize AI workload types and when to use each: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to exam scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: workload selection and governance questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed mini-sim: Describe AI workloads domain sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize AI workload types and when to use each: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to exam scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: workload selection and governance questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed mini-sim: Describe AI workloads domain sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize AI workload types and when to use each: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: AI vs ML vs deep learning—what the exam expects

AI-900 uses “AI” as the umbrella: any system that exhibits human-like capabilities such as perception, language, and decision support. “Machine learning” is a subset of AI where models learn patterns from data rather than being explicitly programmed. “Deep learning” is a further subset of ML that uses multi-layer neural networks, especially common in vision, speech, and modern language models.

On the exam, you’re rarely asked to derive algorithms, but you are expected to classify a scenario as rules-based automation versus ML-based prediction. A common trap is assuming every intelligent behavior requires ML. If the prompt describes deterministic logic (“if customer is in region A and order total exceeds X, route to queue Y”), that’s rules or workflow automation, not ML.

Another trap is equating “deep learning” with “generative AI.” Deep learning also powers classical tasks like image classification and object detection. Generative AI (like large language models) is typically used to create new text, code, or images, or to answer questions in natural language. If the scenario’s output is a label or probability (fraud risk score, churn likelihood), that’s predictive ML; if it’s a paragraph or conversation, you’re in NLP or generative AI territory.

Exam Tip: Look for the output format: numbers and categories suggest ML prediction/classification; bounding boxes suggest detection; extracted fields suggest information extraction; fluent natural language responses suggest conversational AI or generative AI.

Section 2.2: Workload types—prediction, classification, detection, recommendation

This section maps directly to frequent AI-900 scenario stems. “Prediction” typically means forecasting a numeric value: sales next month, delivery time, energy consumption. That aligns to regression. “Classification” means choosing a category: spam vs not spam, high/medium/low risk, product type, sentiment label. “Detection” most often refers to finding something in an image (object detection) or identifying anomalies (anomaly detection) in time series or telemetry. “Recommendation” means ranking items for a user based on behavior, history, or similarity (products, articles, movies).

Azure framing matters. For classic tabular prediction/classification, Azure Machine Learning is the common answer when custom training is implied. For image detection/classification/OCR, Azure AI Vision is the typical managed service when you are not asked to build a bespoke model. For recommendations, the exam expects you to recognize the workload even if the service naming evolves; focus on “recommendations require ranking based on user-item interactions,” not just “classification of products.”

Distractors often swap terms: classification vs detection. If the prompt says “identify whether an image contains a dog,” that’s classification. If it says “locate all dogs in the image and draw rectangles,” that’s object detection. Similarly, recommendation is not the same as prediction: “predict whether a customer will buy item X” is a classification/prediction problem; “show the top 5 items the customer is likely to buy” is recommendation (ranking).

Exam Tip: Under time pressure, reduce each scenario to: input type + output type + action. “Image + bounding boxes + count” → detection. “Tabular + numeric value” → prediction/regression. “User history + ranked list” → recommendation.

Section 2.3: Knowledge mining and information extraction scenarios

Knowledge mining is about turning large volumes of unstructured or semi-structured content into searchable, structured insights. AI-900 commonly frames this as: “We have thousands of PDFs, scans, emails, and need to search, extract key phrases, and find answers.” The correct mental model is an ingestion pipeline: extract text (OCR if needed), enrich with NLP (entities, key phrases, sentiment, language detection), and index for search. On Azure, this is typically anchored by Azure AI Search with AI enrichment.

Information extraction differs from conversational generation. Extraction means pulling specific fields or facts (invoice number, total amount, date, named entities, PII). If the prompt emphasizes compliance, review, or data entry automation, you’re likely in OCR + entity extraction rather than a chatbot. A common trap is choosing a generative AI answer because it “can read documents,” but the exam usually prefers purpose-built extraction when the output is structured fields.

Watch for keywords: “index,” “search across documents,” “enrich content,” “metadata,” “entity recognition,” “OCR,” “forms,” “scanned documents.” Those cue knowledge mining. Another distractor is over-selecting Azure Machine Learning. Unless the prompt says “train a custom model,” managed cognitive services + search are the standard selection.

Exam Tip: If the scenario’s success metric is “find relevant documents quickly” or “enable search,” your anchor is search indexing; extraction/enrichment is supporting. If the success metric is “capture fields accurately,” your anchor is OCR/document processing, not chatbot tech.

Section 2.4: Conversational AI and automation scenarios—bots and copilots

Conversational AI scenarios revolve around interacting with users in natural language through chat or voice, often to answer FAQs, gather information, or trigger actions. AI-900 expects you to distinguish between (1) intent-based bots that route conversations to actions (traditional language understanding) and (2) generative copilots that produce open-ended answers, summaries, or grounded responses using provided content.

Scenario cues: “customer support chat,” “IT helpdesk,” “book an appointment,” “reset password,” “track an order,” “handoff to human agent.” In these, the bot needs orchestration, guardrails, and integration with business systems. If the prompt emphasizes deterministic flows and known intents, think language understanding + bot framework patterns. If it emphasizes drafting responses, summarizing policy, or answering questions from knowledge bases, think Azure OpenAI-style generative responses—ideally with grounding (retrieval) to reduce hallucinations.

Automation is a frequent distractor: some scenarios are simple workflow automation (no AI needed). If the user asks for “create a ticket when email contains ‘urgent’,” that’s pattern matching or rules. If it asks to “understand the user’s request even with varied phrasing,” that pushes you into NLP. Exam Tip: Look for “varied wording,” “natural language,” “intent,” “entities,” and “conversation state” as the tell that AI is required.

Finally, differentiate “bot UI” from “AI brain.” The exam may offer an interface option as a distractor. The correct answer usually targets the capability (language understanding/generative model) rather than the channel (Teams, web chat) unless the prompt explicitly asks for the deployment surface.

Section 2.5: Responsible AI principles—fairness, reliability, privacy, inclusiveness, transparency, accountability

Responsible AI is not a separate topic on AI-900—it’s embedded in workload selection and governance. You’re expected to recognize risk signals in scenario text and map them to principles: fairness (avoid biased outcomes), reliability and safety (consistent behavior, robust performance), privacy and security (protect sensitive data), inclusiveness (works for diverse users and abilities), transparency (users understand when AI is used and why), and accountability (humans remain responsible, with auditability and oversight).

Fairness cues: hiring, lending, insurance pricing, criminal justice, healthcare triage—any decision affecting people’s opportunities. The trap is choosing “maximize accuracy” as the primary goal when the prompt implies equity or regulatory scrutiny. Reliability cues: safety-critical environments, real-time operations, medical devices, autonomous systems. Privacy cues: PII, PHI, financial records, children’s data, or “must comply with regulations.” Inclusiveness cues: accessibility, multilingual audiences, speech impairments, low-bandwidth environments. Transparency cues: “explain why denied,” “users must be informed,” “auditable decisions.” Accountability cues: “human review,” “escalation,” “appeals process,” “monitoring and retraining.”

Exam Tip: When the prompt includes “must be explainable” or “users can appeal,” choose options that include interpretability, documentation, and human-in-the-loop—not only model performance. Another common trap is assuming privacy is solved by “anonymizing” without considering access control, encryption, and data minimization; the exam often expects the broader governance posture.

Section 2.6: Exam-style practice—scenario mapping and eliminating distractors

Your mini-sim performance improves when you apply a repeatable elimination routine. Step 1: classify the input (text, image, speech, tabular, documents). Step 2: classify the output (label/score, extracted fields, bounding boxes, ranked list, natural language response). Step 3: note constraints (real-time vs batch, custom vs prebuilt, regulated vs low-risk). Step 4: pick the workload family and only then match to the service family named in the options.

Eliminating distractors is often faster than proving the correct answer. Remove options that require custom training when the prompt never mentions labeled data or model building. Remove options that are “platform” when the prompt asks for a specific capability (for example, a bot channel vs language analysis). Remove options that mismatch output type (summarization tool for a numeric forecast; OCR tool for sentiment classification). If two options remain, prefer the one that is purpose-built and minimal for the task described.

Time traps in timed sprints: over-reading and second-guessing. Instead, underline (mentally) one noun for input and one verb for output. “Detect defects” on a manufacturing line implies vision detection; “predict maintenance needs” implies ML prediction/anomaly detection; “extract invoice totals” implies OCR + extraction; “answer employee questions from policy documents” implies retrieval-grounded conversational AI.

Exam Tip: After each practice set, do “weak spot repair” by writing a one-line rule for the distractor you fell for (e.g., “classification ≠ detection; detection draws boxes”). This converts mistakes into fast heuristics for the next timed mini-sim.

Chapter milestones
  • Recognize AI workload types and when to use each
  • Apply responsible AI principles to exam scenarios
  • Practice set: workload selection and governance questions
  • Timed mini-sim: Describe AI workloads domain sprint
Chapter quiz

1. A retail company wants to automatically extract the total amount and invoice number from scanned PDF invoices. They prefer a managed service and do not want to build a custom ML pipeline. Which AI workload and Azure service family should you choose?

Show answer
Correct answer: Computer vision (OCR) using Azure AI Vision
This is an OCR/document text extraction scenario (images/PDFs in, text fields out). Azure AI Vision provides OCR capabilities as a managed service. Azure Machine Learning is for building/training custom models (unnecessary when managed OCR meets the need). Azure AI Search is for indexing and searching content; it can enrich content but is not the most direct choice for extracting text from scanned invoices.

2. A support team wants to classify incoming customer emails into categories such as Billing, Technical Issue, and Cancellation. The input is unstructured text, and the output is a label. Which workload best fits?

Show answer
Correct answer: Natural language processing (text classification) using Azure AI Language
The scenario is text-in/label-out, which maps to NLP text classification in Azure AI Language. Azure AI Vision is for images, not email text. Bot services are for building conversational interfaces; they may use language services, but the core workload here is classification rather than managing a dialog.

3. A healthcare provider is considering an AI system to help prioritize patients for follow-up care. The decision could significantly impact individuals. Which responsible AI action is most appropriate to address governance concerns in this scenario?

Show answer
Correct answer: Implement human review for high-impact decisions and document model limitations
For high-impact scenarios, AI-900 emphasizes responsible AI principles such as accountability, transparency, and human-in-the-loop review. Maximizing complexity at the expense of explainability conflicts with transparency and governance requirements. Simply removing demographic fields does not guarantee fairness because bias can exist through proxies and data imbalance; governance requires evaluation, monitoring, and appropriate controls.

4. A manufacturer has sensor readings (temperature, vibration, pressure) and wants to predict whether a machine will fail in the next 7 days. Which workload and Azure service family is the best match?

Show answer
Correct answer: Predictive machine learning using Azure Machine Learning
This is a predictive analytics scenario using tabular/time-series features to output a probability or label (failure/no failure), which aligns with Azure Machine Learning. Object detection is for images with bounding boxes, not sensor tables. Speech-to-text is for audio transcription, unrelated to predicting equipment failure.

5. A company wants employees to type natural-language questions like "What is our vacation policy?" and get answers sourced from internal HR documents. They want a managed approach optimized for searching and retrieving relevant passages from content. Which service family is most appropriate?

Show answer
Correct answer: Azure AI Search (knowledge mining/retrieval over documents)
The requirement is question answering grounded in internal documents, which aligns with retrieval/knowledge mining scenarios typically supported by Azure AI Search (indexing and retrieving relevant content). Azure Machine Learning would be overkill unless the prompt required custom model development and training. Azure AI Vision OCR can extract text but does not provide the core retrieval and ranking needed to answer questions across a document corpus.

Chapter 3: Fundamental Principles of ML on Azure—Train, Evaluate, Deploy

AI-900 expects you to understand machine learning (ML) as a lifecycle, not as a single “train a model” event. On the exam, many questions are designed to see whether you can separate training from inference, pick an appropriate learning approach (supervised/unsupervised/reinforcement), and interpret the most common evaluation metrics without overcomplicating them. In Azure terms, you must also recognize the high-level parts of an ML workflow—what a workspace is, why you need compute, and how experiments and pipelines relate to repeatability and deployment.

This chapter is your weak-spot repair kit for ML fundamentals: you’ll practice translating business statements into ML problem types, predicting which metric matters, and spotting the common traps (like treating accuracy as universally “best,” or mixing up validation and test data). You are not expected to build deep models from scratch in AI-900; you are expected to choose sensible options and explain what the numbers mean.

Exam Tip: When an answer choice sounds like it requires advanced math or detailed algorithm tuning, it’s usually not the AI-900 target. Prefer answers focused on correct problem framing, dataset splits, and plain-language metric interpretation.

Practice note for Choose the right ML approach: supervised, unsupervised, and reinforcement learning: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Interpret core ML concepts: features, labels, overfitting, and metrics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Select Azure tooling for ML workflows (conceptual, exam-level): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: ML fundamentals and Azure ML scenario questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right ML approach: supervised, unsupervised, and reinforcement learning: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Interpret core ML concepts: features, labels, overfitting, and metrics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Select Azure tooling for ML workflows (conceptual, exam-level): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: ML fundamentals and Azure ML scenario questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right ML approach: supervised, unsupervised, and reinforcement learning: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Interpret core ML concepts: features, labels, overfitting, and metrics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: ML lifecycle—data prep, training, validation, testing, inference

AI-900 tests the ML lifecycle as a sequence of distinct phases with distinct goals: data preparation, training, validation, testing, and inference (deployment/use). The exam frequently uses scenarios where a team “trained a model and it did great” but then fails in production—your job is to identify which phase was skipped or misused.

Data prep includes collecting data, cleaning (missing values, duplicates), encoding categorical values, and feature engineering (creating inputs that help the model learn). A common exam angle is that “better data” often beats “more complex algorithm.” If the prompt mentions inconsistent formats or lots of nulls, the correct action is typically data preparation, not changing the model type.

Training is when the algorithm learns patterns from labeled or unlabeled data. In supervised learning, training uses features (inputs) and labels (the target). Inference is separate: it’s the act of using the trained model to make predictions on new data.

Validation is used during model development to tune choices (hyperparameters, feature selection, thresholds) without “peeking” at the test set. Testing is the final, unbiased estimate of performance on unseen data—ideally representative of production.

  • Training set: fit model parameters.
  • Validation set: choose settings and compare candidate models.
  • Test set: report final expected performance.
  • Inference endpoint: where production predictions happen (batch or real-time).

Common trap: Using the test set to tune the model. If the scenario implies repeated evaluations on the test data while adjusting the model, that’s leakage; the test metric becomes optimistic.

Exam Tip: If you see “model performs well in the lab but poorly after deployment,” think about data drift (production data differs), leakage, or an unrepresentative test set. The correct fix is often to retrain with newer data, monitor, and validate on data resembling production.

Section 3.2: Learning types—supervised vs unsupervised vs reinforcement learning

Choosing the right ML approach is a core exam objective. AI-900 keeps this conceptual: identify which learning type fits the data you have and the outcome you need.

Supervised learning applies when you have labeled examples (inputs with known outputs). Typical tasks: classification (predict a category) and regression (predict a numeric value). If the scenario mentions “historical records with outcomes,” that’s your cue for supervised learning.

Unsupervised learning applies when you do not have labels and want to discover structure, often via clustering. If the prompt says “group customers by similar behavior” or “find segments,” that points to unsupervised learning.

Reinforcement learning is about an agent learning actions through rewards and penalties, often in sequential decision-making (robot navigation, game playing, dynamic resource allocation). AI-900 usually frames it as “learn the best action over time” rather than “predict a label.”

  • Supervised: labeled data → predict label/value.
  • Unsupervised: unlabeled data → find patterns/groups.
  • Reinforcement: reward signal → optimize behavior through trial and error.

Common trap: Mistaking anomaly detection or segmentation as supervised. Unless the scenario clearly states labeled “anomaly” examples, grouping-like tasks are typically unsupervised (clustering). Another trap is assuming reinforcement learning whenever “optimize” appears; many optimization problems are still supervised (predict demand) followed by business rules.

Exam Tip: Anchor on the presence (or absence) of labels. If the question gives you a column like “Churn: Yes/No,” it’s almost certainly supervised classification.

Section 3.3: Problem framing—classification vs regression vs clustering

After choosing supervised vs unsupervised, the next exam skill is correct problem framing. AI-900 loves subtle wording: “predict whether” vs “estimate how much” vs “group similar.” Map each to the right technique.

Classification predicts discrete categories (fraud vs not fraud, spam vs not spam, approved/denied). Binary classification is common, but multi-class also appears (categorize support tickets into departments). If outputs are categories, it’s classification.

Regression predicts a continuous numeric value (house price, demand forecast, time-to-failure). If the question asks for an “amount,” “score,” “temperature,” or “number of units,” regression is typically correct.

Clustering groups items into clusters based on similarity without pre-defined labels (customer segmentation, grouping IoT devices by behavior). Outputs are cluster assignments discovered from data rather than “correct” labels.

  • Key words for classification: yes/no, type, category, class, label.
  • Key words for regression: estimate, forecast, value, amount, continuous.
  • Key words for clustering: segment, group, similar, discover patterns.

Common trap: Confusing “rank” with regression. Ranking (e.g., ordering recommendations) can involve regression-like scores, but AI-900 usually expects you to pick classification/regression/clustering based on the described output. If the output is a probability of purchase, that’s often classification (will buy vs won’t) even though the model outputs a probability.

Exam Tip: Ignore the input data type (text, images, numeric) when choosing classification vs regression. The question is about the output you need, not the modality.

Section 3.4: Model evaluation—accuracy, precision/recall, RMSE, confusion matrix basics

AI-900 does not require you to compute metrics by hand, but it does require you to choose and interpret them. The exam frequently contrasts accuracy with precision/recall, and expects you to know that regression uses different measures like RMSE.

Accuracy is the fraction of correct predictions. It can be misleading with imbalanced data (e.g., 99% non-fraud). If almost everything is the negative class, predicting “not fraud” always yields high accuracy but a useless model.

Precision answers: “When the model predicts positive, how often is it correct?” Precision matters when false positives are costly (flagging legitimate transactions as fraud, or incorrectly diagnosing disease).

Recall answers: “Of all actual positives, how many did we catch?” Recall matters when missing positives is costly (failing to detect fraud, missing a safety issue). Many scenarios are about trading off precision vs recall via a threshold.

Confusion matrix summarizes true positives, false positives, true negatives, and false negatives. You don’t need to memorize every formula, but you must identify which error type the business cares about.

RMSE (Root Mean Squared Error) measures typical prediction error magnitude in regression. Lower RMSE is better; it’s in the same unit as the target variable (e.g., dollars). It penalizes large errors more than small ones.

Common trap: Choosing accuracy for fraud, rare disease, intrusion detection, or defect detection. In imbalanced classification, precision/recall (and sometimes F1) are usually more informative.

Exam Tip: If the scenario says “minimize missed detections,” choose recall-focused thinking. If it says “avoid false alarms,” think precision. If it predicts a number, look for RMSE (or other regression metrics) rather than confusion matrices.

Section 3.5: Azure ML concepts—workspace, compute, experiments, pipelines (high level)

For AI-900, Azure Machine Learning (Azure ML) is tested at a recognition level: know the major building blocks and what they’re for. Don’t over-index on implementation details (SDK calls, YAML syntax); focus on concepts you can map to scenario questions.

Workspace is the top-level container for Azure ML assets: datasets/data connections, models, runs, endpoints, and linked resources. If a question asks where you manage models, runs, and deployments, “workspace” is the anchor.

Compute refers to the compute targets used for training or inference. Training compute might be a CPU/GPU cluster; inference compute might be a managed endpoint. The exam often checks that you understand training is resource-intensive and can be scaled, and inference can be real-time or batch.

Experiments are organized collections of runs. A “run” is a single execution of a training script/pipeline step with tracked parameters and metrics. This supports reproducibility and comparison (e.g., which model version performed best).

Pipelines represent repeatable workflows (data prep → train → evaluate → register → deploy). Pipelines help operationalize ML and reduce “works on my machine” issues.

  • Workspace: central management and governance boundary.
  • Compute: where training/inference happens.
  • Experiment/run: track metrics, compare iterations.
  • Pipeline: automate and standardize the lifecycle.

Common trap: Mixing Azure ML with Azure AI services. Azure AI services (prebuilt) are often used when you want no/low training; Azure ML is used when you train custom models. If the scenario emphasizes custom training with your own labeled data and model selection, Azure ML is typically the better fit.

Exam Tip: When you see “repeatable,” “automated,” “orchestrate steps,” or “MLOps,” pipelines are a strong clue. When you see “track runs and metrics,” think experiments.

Section 3.6: Exam-style practice—common pitfalls and metric-based choices

This chapter’s skills come together in the exam’s scenario style. Your success depends less on memorizing definitions and more on a consistent decision process: identify the goal, identify labels, select learning type, select problem framing, then select an evaluation metric that matches business risk.

Metric-based choices: If the prompt includes rare events (fraud, defects, security incidents), treat “high accuracy” skeptically. Ask yourself which is worse: a false positive (unnecessary escalation) or a false negative (missed incident). That directly maps to precision vs recall. For numerical forecasts, select regression metrics like RMSE and look for language about “average error magnitude.”

Overfitting signals: The exam may describe “excellent training performance but poor test performance.” That’s classic overfitting. The correct remedies at AI-900 level are conceptual: more data, simpler model, regularization, or better validation—rather than obscure algorithm tweaks.

Training vs inference trap: If a question describes “using the model to make predictions on new images/text,” that’s inference. If it describes “learning from labeled examples,” that’s training. Many wrong options swap these terms.

Azure workflow traps: If the scenario emphasizes rapid adoption with minimal ML expertise and prebuilt capabilities, that leans toward Azure AI services. If it emphasizes custom features, custom labels, and experiment tracking, it leans toward Azure ML workspace/experiments/pipelines.

Exam Tip: When two answers both sound plausible, choose the one that aligns with the stated business cost of errors and the presence/absence of labels. These two clues usually eliminate distractors faster than technical jargon does.

Chapter milestones
  • Choose the right ML approach: supervised, unsupervised, and reinforcement learning
  • Interpret core ML concepts: features, labels, overfitting, and metrics
  • Select Azure tooling for ML workflows (conceptual, exam-level)
  • Practice set: ML fundamentals and Azure ML scenario questions
Chapter quiz

1. A retail company wants to predict the next-day demand (number of units) for each product based on historical sales, promotions, and weather. Which machine learning approach should you choose?

Show answer
Correct answer: Supervised learning (regression)
Predicting a numeric value from labeled historical examples is supervised learning (regression). Clustering is unsupervised and groups similar items without a target label, so it won’t directly predict demand. Reinforcement learning focuses on learning actions via rewards in an environment (e.g., robotics or dynamic decision policies), not straightforward numeric prediction from labeled data.

2. You trained a binary classification model to detect fraudulent transactions. Fraud cases are rare (about 1% of transactions). Which evaluation metric is typically most useful to understand how many flagged transactions are actually fraud and how many frauds you catch?

Show answer
Correct answer: Precision and recall
With imbalanced classes, precision and recall provide meaningful insight: precision indicates how many predicted frauds are truly fraud, and recall indicates how many true frauds are detected. Accuracy can be misleading (a model that always predicts 'not fraud' could be ~99% accurate). R-squared is a regression metric and does not apply to binary classification.

3. A data scientist says, "Our model performs extremely well on the training dataset but significantly worse on the validation dataset." What is the most likely issue?

Show answer
Correct answer: Overfitting
A large gap between training performance and validation performance commonly indicates overfitting (the model memorizes training patterns and fails to generalize). Underfitting would usually result in poor performance on both training and validation. The learning type (supervised vs. unsupervised) isn’t inferred from this symptom; the issue described is about generalization across splits.

4. You are designing an ML workflow in Azure. You want a central place to manage datasets, experiments, models, and deployments for a team. Which Azure resource is primarily used for this purpose?

Show answer
Correct answer: Azure Machine Learning workspace
An Azure Machine Learning workspace is the top-level resource for organizing and managing ML assets (experiments, data references, registered models, endpoints) at an exam-concept level. Blob Storage stores data but does not manage experiments/models/deployment lifecycle. AKS is a compute platform often used to host deployed models, but it is not the central ML management resource.

5. A team has trained a model in Azure and wants to use it to generate predictions in a business application. Which statement best describes the difference between training and inference?

Show answer
Correct answer: Training learns model parameters from labeled data; inference uses the trained model to make predictions on new data.
Training is the learning phase where the algorithm fits parameters based on data (typically including labels for supervised learning). Inference is the prediction/scoring phase on new, unseen data. Deploying as an endpoint is part of operationalization, not training itself, and evaluating on the training set is not inference. Both supervised and unsupervised models can have an inference/scoring step (e.g., assigning clusters), so restricting them as in option C is incorrect.

Chapter 4: Computer Vision Workloads on Azure—Pick the Right Service

AI-900 rarely tests whether you can build a full vision pipeline; it tests whether you can recognize the type of vision problem, the shape of the output (label, bounding box, text, embedding), and then pick the right Azure service. In this chapter you’ll practice that exam reflex: read a scenario, identify the core vision task, and map it to the service that produces the needed output with the least custom work.

The most common confusion on the exam is mixing up “image analysis” (understanding what’s in a picture) with “object detection” (where an object is) and “OCR” (what text is visible). Another frequent trap: choosing a training service (custom model building) when the prompt clearly wants an out-of-the-box API, or choosing a generic service when the scenario demands domain-specific extraction.

We will follow the chapter lessons in sequence: (1) identify core computer vision tasks and outputs, (2) map scenarios to Azure vision services and capabilities, (3) practice set themes around OCR/image analysis/detection, and (4) finish with a timed mini-sim mindset—how to sprint through service-selection questions without overthinking.

  • Core tasks: classification vs detection vs segmentation vs OCR
  • Core services: Azure AI Vision (Image Analysis, OCR/Read), Azure AI Face (where permitted), Azure AI Custom Vision, and Azure Machine Learning for fully custom approaches
  • Outputs you must recognize: tags/captions, bounding boxes, polygons/masks, extracted text, confidence scores

Exam Tip: When a question describes “identify what is in the image” and does not mention training, default to a prebuilt vision API. When it describes “find each instance and its location,” you’re in detection territory. When it describes “extract printed/handwritten text,” it’s OCR. Force yourself to say the output out loud before picking the service.

Practice note for Identify core computer vision tasks and outputs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map scenarios to Azure vision services and capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: OCR, image analysis, and detection questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed mini-sim: computer vision domain sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify core computer vision tasks and outputs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map scenarios to Azure vision services and capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: OCR, image analysis, and detection questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed mini-sim: computer vision domain sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Vision workload overview—classification, detection, segmentation (conceptual)

Section 4.1: Vision workload overview—classification, detection, segmentation (conceptual)

Computer vision questions on AI-900 start with a deceptively simple skill: naming the workload. The exam expects you to distinguish between classification, object detection, and segmentation based on the wording and the required output.

Classification answers “What is this image?” or “Which category best describes it?” The output is typically a label (or multiple labels) with confidence scores. This is used for sorting, routing, or moderation-like categorization. Object detection answers “Where are the objects?” and returns bounding boxes (x/y/width/height) with a class label and confidence for each detected instance. Segmentation goes further: it identifies the exact pixel region of an object (a mask/polygon), which is useful when shape matters (defects, medical imaging, or background removal).

AI-900 often embeds the answer in the output requirement: if you see “bounding box” or “coordinates,” think detection; if you see “mask” or “pixels,” think segmentation; if you see “label only,” think classification. This is part of “Identify core computer vision tasks and outputs.”

Common trap: confusing detection with classification when the scenario says “identify all items on a shelf.” “All items” implies multiple instances and location, which points to detection, not single-label classification. Another trap is assuming every vision scenario needs training—many exam prompts intend prebuilt analysis.

Exam Tip: In timed questions, do a two-step parse: (1) What is the user asking for—label, location, or pixels? (2) Do they mention a need to train on company-specific images? Only then consider custom models.

Section 4.2: Image analysis scenarios—tags, captions, and content understanding

Section 4.2: Image analysis scenarios—tags, captions, and content understanding

Image analysis scenarios on AI-900 are the “general understanding” bucket: generate tags, captions, detect common objects/attributes, and return a structured description of what’s present. In Azure, this is typically framed as using Azure AI Vision image analysis capabilities (often described as extracting tags, descriptions/captions, categories, or detecting adult/racy content depending on the wording of the exam item).

What the exam tests is your ability to match non-location understanding tasks to the right service. If the scenario says “add alt text for accessibility,” “auto-caption images,” “tag photos for search,” or “summarize what’s in the picture,” think prebuilt image analysis. The output is usually a list of tags/objects with confidence scores and possibly a natural-language caption. This differs from detection because the prompt often doesn’t require precise coordinates.

Common trap: choosing OCR when the scenario says “understand a meme” or “caption a photo.” Even if text may exist, the primary requirement is semantic understanding (caption/tags), not text extraction. Another trap is picking Custom Vision simply because the organization has images; unless the prompt explicitly calls for training on proprietary classes, prebuilt image analysis is the expected exam answer.

How to identify correct answers: Look for verbs like “describe,” “tag,” “categorize,” “generate a caption,” “identify landmarks,” or “detect inappropriate content.” These map to image analysis. If the question mentions “return JSON with tags and confidence,” that’s also a strong signal.

Exam Tip: When multiple services look plausible, prefer the one that is (a) prebuilt and (b) directly produces the requested output type. AI-900 rewards minimal-sufficient service selection, not architecture ambition.

Section 4.3: OCR and document extraction—text in images and forms-style use cases

Section 4.3: OCR and document extraction—text in images and forms-style use cases

OCR is one of the highest-frequency vision objectives: extracting printed or handwritten text from images and documents. The key phrase the exam wants you to notice is “text in an image,” “scan,” “receipt,” “invoice,” “ID card,” “PDF,” “handwritten notes,” or “convert images to editable text.” In Azure service terms for AI-900, expect Azure AI Vision OCR/Read concepts (often described simply as an OCR capability that returns lines/words with positions).

OCR outputs differ from image analysis outputs. OCR returns the extracted characters plus structure: lines, words, and often bounding boxes around words/lines, along with confidence. This “where the text is” is not object detection; it is text layout information. The exam may include hints like “store the extracted text,” “index the text,” or “enable search over scanned documents”—all OCR-aligned.

Forms-style use cases add a subtle layer: the goal is not just reading text, but extracting fields (invoice number, total, date) and sometimes tables. On AI-900, you’re usually not expected to design a complex document intelligence solution, but you should recognize that a scenario describing key-value extraction from forms is beyond generic tagging/captioning. If the prompt emphasizes structured fields from business documents, think document extraction capabilities rather than plain image captions.

Common trap: selecting image analysis because “it’s an image,” even though the requirement is “extract text.” Another trap is confusing OCR with translation: OCR gets you text; translation is a subsequent NLP step.

Exam Tip: If the scenario includes “handwritten” or “scanned PDF,” OCR/Read is the anchor service. Then ask: do they need raw text + coordinates (OCR) or business fields (forms/invoice-like extraction)? Pick the most direct match.

Section 4.4: Face and person-related scenarios—capabilities, constraints, and ethics

Section 4.4: Face and person-related scenarios—capabilities, constraints, and ethics

Face/person scenarios are where AI-900 blends service selection with Responsible AI. The exam may describe: detecting whether a face is present, locating faces in an image, or analyzing facial attributes in allowed contexts. In Azure, these map to Azure AI Face capabilities when the requirement is explicitly about faces (not general objects). The typical outputs are face bounding boxes (location) and face-related metadata depending on what’s permitted.

However, the exam also expects you to know there are constraints and ethical considerations around facial recognition and sensitive attribute inference. If the scenario implies identifying a specific person (“who is this?”) or performing surveillance-like tracking, this triggers Responsible AI red flags and service policy considerations. In many exam-style prompts, the “best answer” is the one that avoids inappropriate biometric identification and instead uses less intrusive alternatives (for example, detecting presence of people without identifying individuals).

Common trap: assuming “face detection” and “face recognition” are the same. Detection is “a face exists here” (location). Recognition/verification is “this is the same person” or “match to an identity.” If the question only needs “count how many people are in the photo” or “blur faces,” it’s detection/localization, not identity.

Exam Tip: When you see person-related scenarios, pause and check for policy triggers: identification, tracking across cameras, or inferring sensitive traits. On AI-900, demonstrating Responsible AI awareness can be the differentiator between two plausible choices.

How to identify correct answers: If the prompt says “detect faces” or “return face rectangles,” Face is likely. If it says “detect objects/people” without face specificity, general vision detection may be sufficient. If it asks to identify an employee by face, expect the exam to steer you toward compliance/ethics constraints rather than a purely technical selection.

Section 4.5: Custom vs prebuilt—when to use out-of-box vs custom vision concepts

Section 4.5: Custom vs prebuilt—when to use out-of-box vs custom vision concepts

The exam loves “custom vs prebuilt” because it tests judgment, not memorization. Prebuilt vision works when your task matches common concepts (tags, captions, OCR, generic detection). Custom vision is for when your organization has domain-specific classes the prebuilt model won’t reliably recognize (defects on a manufactured part, specific product SKUs, proprietary symbols, unusual safety gear types).

In Azure, custom model building is commonly associated with Azure AI Custom Vision concepts: you provide labeled images, train a model, and then run inference. The key exam indicator is the requirement for training on your images: “train,” “label,” “improve accuracy for our products,” “new categories,” “factory-specific defects,” or “detect our brand’s packaging variations.” Those phrases are signals that prebuilt image analysis will be insufficient.

Common trap: picking custom vision just because the scenario mentions “high accuracy.” The exam usually expects prebuilt services unless there’s explicit evidence that the classes are unique or that a custom dataset exists and training is feasible. Another trap is ignoring the operational split: training happens once in a controlled phase; inference is the runtime prediction. AI-900 can ask which phase needs labeled data (training) versus which phase runs on new images (inference).

Exam Tip: Use the “3 U’s” test for custom vision: Unusual classes, Unique environment, Unmet by prebuilt. If at least one is clearly stated, custom becomes likely. If none are stated, default to prebuilt to match the exam’s intent.

Finally, remember output alignment: classification models output labels; detection models output bounding boxes. Custom Vision can be used for both classification and detection, and the scenario’s output requirement should dictate which you choose.

Section 4.6: Exam-style practice—service selection and output interpretation

Section 4.6: Exam-style practice—service selection and output interpretation

This section mirrors the chapter’s “practice set” and “timed mini-sim” goals without turning into a quiz. Your job in the exam sprint is to (1) spot the workload, (2) choose the service, and (3) verify the output matches what the scenario needs. Most wrong answers fail step (3): they pick a plausible service but the output type doesn’t satisfy the requirement.

Service selection patterns to memorize:

  • If the output is tags/captions for search/accessibility: choose prebuilt Azure AI Vision image analysis.
  • If the output is text extraction from images/PDFs with word/line locations: choose OCR/Read capabilities in Azure AI Vision.
  • If the output is bounding boxes for objects (multiple instances): choose an object detection capability; if classes are custom, consider Azure AI Custom Vision detection.
  • If the output is face rectangles or face-specific analysis (within policy constraints): consider Azure AI Face, and check Responsible AI implications.

Output interpretation: AI-900 may present short excerpts of results (confidence scores, rectangles, lists of words). Train yourself to map outputs back to tasks. Lists of “tag: confidence” imply classification/image analysis. Arrays of “rectangle: x,y,w,h” imply detection/localization. Hierarchies of “page → line → word” imply OCR. If you see confidence values, remember they are not “accuracy”; they are model certainty for that specific prediction.

Common timed-sim trap: over-reading the scenario and inventing requirements. The exam text is usually sufficient: if it never says “train,” don’t choose a training-centric solution. If it never says “identify a person,” don’t assume recognition is required. Answer what is asked, not what might be nice.

Exam Tip: In a domain sprint, spend your first pass eliminating answers that produce the wrong output type. Only then compare the remaining services on whether they are prebuilt vs custom and whether Responsible AI constraints make one choice unsafe or noncompliant.

Chapter milestones
  • Identify core computer vision tasks and outputs
  • Map scenarios to Azure vision services and capabilities
  • Practice set: OCR, image analysis, and detection questions
  • Timed mini-sim: computer vision domain sprint
Chapter quiz

1. A retail company wants an application that returns a short description of what is happening in a product photo (for example, “a person holding a red backpack”) and a set of tags (for example, “person”, “backpack”, “indoor”). The company does not want to train a model. Which Azure service and feature should you use?

Show answer
Correct answer: Azure AI Vision Image Analysis
Azure AI Vision Image Analysis provides prebuilt captioning and tagging for “what is in the image” without requiring training. Custom Vision classification is for training your own classifier (unnecessary when the prompt asks for out-of-the-box results). Azure Machine Learning is a fully custom approach and is overkill when a prebuilt vision API meets the requirement.

2. A company is building a safety solution that must locate every hard hat in images taken at a construction site and return the coordinates for each hat. No custom training is requested. Which computer vision task and output best match the requirement?

Show answer
Correct answer: Object detection that returns bounding boxes and confidence scores
The requirement is to find each instance and its location, which is object detection, typically returning bounding boxes (and confidence). Classification returns a single label for the entire image and does not provide location. OCR extracts text and is unrelated to locating physical objects like hard hats.

3. An insurance company must extract the policy number and customer name from scanned claim forms that may include handwritten fields. The solution should return the recognized text. Which Azure service capability should you use?

Show answer
Correct answer: Azure AI Vision OCR (Read) to extract printed and handwritten text
Azure AI Vision OCR (Read) is designed to extract printed and handwritten text from images and scanned documents. Image Analysis focuses on describing image content (tags/captions) rather than returning text. Custom Vision object detection can locate regions but does not inherently perform OCR; it would add unnecessary training and still would not directly return the recognized characters.

4. A manufacturer wants to identify defects on a specific part type. The defects are unique to their product line, and the company has labeled images for training. They want the model to return the location of each defect in an image. Which service should you choose?

Show answer
Correct answer: Azure AI Custom Vision object detection
Custom Vision object detection is appropriate when you must train on your own labeled dataset and need localized results (bounding boxes) for each instance (defect). Prebuilt Image Analysis is intended for general-purpose understanding and typically won’t detect product-specific defect classes without training. OCR (Read) is for extracting text, not detecting visual defects.

5. You are answering an AI-900-style question: “You need to process product images and return where the company logo appears in each image.” The prompt does not mention training. What is the best first step to select the correct service?

Show answer
Correct answer: Determine the expected output type (for example, bounding box vs text vs tags) before choosing a service
AI-900 service-selection questions commonly hinge on recognizing the output shape: logo location implies detection outputs (for example, bounding boxes). Azure Machine Learning is not the default when an out-of-the-box vision API can satisfy the scenario and the question does not ask for custom training. OCR is only appropriate when the requirement is to extract readable text, not to locate a logo as an object.

Chapter 5: NLP + Generative AI Workloads on Azure—Language to LLMs

This chapter targets the AI-900 objective area that asks you to recognize NLP and generative AI workloads and select the right Azure service for each scenario. In the Mock Exam Marathon format, you’re not trying to become a language engineer—you’re trying to quickly map a business ask (“analyze reviews,” “build a support chatbot,” “translate calls,” “generate a draft email”) to the correct Azure offering and a few key concepts the exam expects you to know. The chapter is organized to match the lesson flow: first, choosing Azure NLP services for analytics, translation, and conversations; then, understanding generative AI basics (LLMs, prompts, grounding, safety); and finally, preparing you for mixed question sets and a timed mini-sim where speed and pattern recognition matter.

On AI-900, many wrong answers are “almost right” services. Your job is to spot the keywords that anchor the workload type: extract (key phrases/entities), classify (sentiment, language, intent), converse (bots), translate (Text Translation), or generate (Azure OpenAI). As you read, practice turning each requirement into: (1) workload type, (2) likely Azure service family, and (3) any responsible AI or safety constraint that should be mentioned.

Practice note for Choose Azure NLP services for analytics, translation, and conversations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand generative AI basics: LLMs, prompts, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: NLP and generative AI mixed questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed mini-sim: language and generative AI domain sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose Azure NLP services for analytics, translation, and conversations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand generative AI basics: LLMs, prompts, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: NLP and generative AI mixed questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Timed mini-sim: language and generative AI domain sprint: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose Azure NLP services for analytics, translation, and conversations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand generative AI basics: LLMs, prompts, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: NLP workload overview—entities, sentiment, key phrases, summarization concepts

NLP “analytics” workloads focus on understanding text rather than generating it. AI-900 commonly tests whether you can identify tasks like entity recognition, sentiment analysis, key phrase extraction, language detection, and summarization, and associate them with Azure’s language services (often referenced as Azure AI Language capabilities). The mental model: you submit text, the service returns structured outputs—labels, scores, spans, and summaries—used downstream in dashboards, routing rules, or search.

Entities are real-world “things” pulled from text (people, organizations, locations, dates). Expect scenarios like “extract company names from news articles” or “detect mentions of medications in notes.” Sentiment analysis classifies opinion (positive/neutral/negative) and may include a confidence score. Key phrase extraction highlights the most important terms—useful for tagging, indexing, and quick insights on large document sets. Summarization (extractive or abstractive, depending on the feature set described) appears when the requirement is “reduce long text to key points” for analysts or agents.

Exam Tip: Watch for the verb in the prompt. If the user wants to pull structure out of text (entities, phrases, sentiment), it’s NLP analytics. If they want to create new text (draft, rewrite, brainstorm), that’s generative AI (Azure OpenAI).

  • Common trap: Picking a bot framework or LLM for a simple extraction task. If the question says “identify key phrases” or “detect sentiment,” don’t overcomplicate it.
  • Common trap: Confusing summarization with search. Summarization produces a condensed version; search retrieves documents.

How to identify correct answers quickly: look for outputs described as “scores,” “labels,” “detected entities,” or “extracted phrases.” Those are classic Language/Text Analytics patterns. In the practice set portion of this chapter, build speed by translating each requirement into one of these standard outputs before you ever look at the choices.

Section 5.2: Language understanding and conversational AI—intents, entities, and bot scenarios

Conversation scenarios on AI-900 often hinge on the difference between recognizing user goals and responding like a human. Language understanding (LU) is about mapping an utterance to an intent (the user’s goal) and extracting entities (parameters needed to fulfill it). For example: “Book a flight to Seattle tomorrow” might map to intent “BookFlight” and entities {destination: Seattle, date: tomorrow}. These are the core exam terms: intent = action; entity = detail.

Conversational AI commonly layers: (1) a channel (web chat, Teams, voice), (2) a bot to manage dialog, and (3) language understanding or a generative model to interpret messages. AI-900 questions frequently describe a support chatbot that must route tickets, look up order status, or capture structured fields. If the scenario emphasizes consistent, predictable routing and slot-filling (collecting required entities), think “intent/entity recognition” style solutions rather than open-ended generation.

Exam Tip: If the requirements include “trigger workflows,” “route to departments,” “fill in a form,” or “ensure consistent answers,” that’s a strong signal for intent/entity-based conversational design. Generative AI can assist, but the exam usually wants you to recognize when classic NLU fits best.

  • Common trap: Selecting translation when the real need is intent recognition. “Supports multiple languages” might be secondary; first decide whether the key need is understanding, translating, or generating.
  • Common trap: Treating a bot as the “AI.” A bot is orchestration; the intelligence comes from language understanding, retrieval, or an LLM.

In the timed mini-sim, you’ll often see a brief transcript-like prompt. Under time pressure, underline the words that imply intent (“I want to…”, “help me…”, “how do I…”) and the fields that must be extracted (account number, product name, date). That points you to intent/entity language understanding patterns.

Section 5.3: Translation and speech intersections—common exam scenario patterns

Translation questions are typically straightforward if you spot the input and output modalities. Text-to-text translation points to Azure translation capabilities (often framed as Azure AI Translator). The exam also mixes in speech scenarios: converting speech to text, translating spoken conversations, or generating synthesized speech. The key is to identify whether the scenario is about language conversion (translation) or audio modality (speech recognition/synthesis) or both.

Common patterns include: “Translate customer emails into English,” “Translate a website UI,” “Provide real-time translation for a call center,” or “Add captions/subtitles in another language.” For call centers and meetings, expect a chain: speech-to-text (recognition), translation, and possibly text-to-speech (voice output). Even when the question doesn’t name the services, the workflow clues are usually explicit: microphone input implies speech; multilingual output implies translation.

Exam Tip: Separate the pipeline into steps. If the prompt starts with audio, your first step is speech recognition. If it ends with spoken output, add speech synthesis. Translation sits in the middle when the language changes.

  • Common trap: Confusing “transcription” with “translation.” Transcription keeps the same language; translation changes it.
  • Common trap: Assuming generative AI is required. If the ask is “translate accurately,” a translation service is more appropriate than an LLM that may paraphrase.

When you face mixed NLP/generative options in the practice set, use an “accuracy vs creativity” test: translation tasks demand faithful conversion (accuracy), while generative tasks are tolerant of variation (creativity). That one heuristic eliminates many distractors quickly.

Section 5.4: Generative AI concepts—tokens, context windows, embeddings (fundamentals)

Generative AI on AI-900 is concept-heavy: you’re expected to understand core terms well enough to select correct statements and match use cases. Start with tokens: models process text in token units (pieces of words/characters). Cost and limits are often token-based, and prompts plus responses both consume tokens. Next, context window: the maximum number of tokens the model can “see” at once (prompt + conversation history + retrieved content). If a scenario mentions “the model forgets earlier messages” or “long documents don’t fit,” the concept being tested is context window limits.

Embeddings are numeric vector representations of text (or other data) that capture semantic meaning. Embeddings are central to semantic search and retrieval-augmented generation (RAG): instead of asking the model to memorize your documents, you store embeddings for your content, retrieve the most relevant passages for a query, and include them as grounding context in the prompt. AI-900 won’t require you to implement RAG, but it may test the idea that embeddings enable similarity matching (e.g., “find the most similar support article”).

Exam Tip: If the scenario says “find similar documents,” “semantic search,” or “retrieve relevant passages,” think embeddings. If it says “generate an answer using company docs,” think grounding/RAG rather than training a new base model.

  • Common trap: Equating grounding with training. Grounding supplies external context at inference time; training changes model weights.
  • Common trap: Ignoring token limits. Long PDFs or multi-turn chats can exceed the context window; the correct answer may mention summarization, chunking, or retrieval.

In the chapter’s weak spot repair drills, practice labeling each generative scenario with: tokens (cost/limits), context window (memory), and embeddings (retrieval/similarity). Those three terms show up repeatedly in AI-900 generative AI items.

Section 5.5: Azure OpenAI scenarios—chat, completions, image generation use cases and constraints

Azure OpenAI Service is the Azure-hosted offering for using OpenAI models with enterprise controls. AI-900 typically tests what you would use it for (drafting, summarizing, Q&A, code assistance, content generation) and what constraints apply (responsible use, content filtering, data handling considerations). You should recognize three common interaction patterns: chat (multi-turn conversation), completions (single-turn prompt-to-output generation), and image generation (creating images from text prompts). The exam often describes these without naming them explicitly.

Chat is best when the user expects a conversational experience (support agent copilot, internal HR assistant, interactive tutoring). Completions fit “generate a product description” or “rewrite this paragraph” style tasks. Image generation is for marketing mockups, concept art, or creating variations—expect the question to mention “create images based on text descriptions.”

Exam Tip: When choices include both “Azure AI Language” and “Azure OpenAI,” decide if the output is structured insights (Language) or newly generated content (OpenAI). If the output is a paragraph, email, or creative text, it’s usually OpenAI.

  • Common trap: Selecting OpenAI for deterministic classification. If the requirement is “classify sentiment” or “extract entities,” Azure NLP analytics is the cleaner fit.
  • Common trap: Assuming OpenAI automatically has your company data. Without grounding (retrieval) or explicit context in the prompt, the model won’t know your internal policies or latest documents.

For the timed mini-sim “domain sprint,” prioritize identifying the interaction style first (chat vs single-turn generation vs images). Then check for constraints like “must cite company policy” (grounding needed) or “must avoid disallowed content” (safety controls).

Section 5.6: Safety and responsible generative AI—content filters, privacy, and human-in-the-loop

AI-900 includes responsible AI expectations across all workloads, but generative AI amplifies them because outputs can be unpredictable. The exam commonly tests that you understand the role of content filters (to reduce harmful, adult, violent, or hateful outputs), privacy/data handling (protecting sensitive inputs), and human-in-the-loop review for high-impact scenarios. Safety isn’t an add-on; it’s part of choosing and operating the solution.

Content filtering helps enforce usage policies and reduce unsafe generations, but it does not guarantee perfection. Privacy considerations include minimizing sensitive data in prompts, access control, auditing, and understanding where data is processed and stored in your Azure environment. Human-in-the-loop is the operational pattern where a person reviews or approves outputs—common for legal, healthcare, finance, or any scenario where mistakes have consequences.

Exam Tip: If a scenario mentions “must not generate offensive content,” “regulatory compliance,” “PII,” or “review before sending,” the correct answer usually includes safety controls and/or human oversight, not just “use an LLM.”

  • Common trap: Treating safety as only a model setting. Real solutions combine filtering, prompt design, grounding to trusted sources, logging/monitoring, and escalation paths.
  • Common trap: Forgetting that prompts can contain sensitive data. The question may test that you should avoid sending unnecessary personal data and apply least-privilege access.

In your mixed practice set review, mark any missed items that involved safety language. Most learners miss these not because they don’t know the concept, but because they focus on functionality first. On AI-900, a “best answer” often balances capability with responsible deployment basics.

Chapter milestones
  • Choose Azure NLP services for analytics, translation, and conversations
  • Understand generative AI basics: LLMs, prompts, grounding, and safety
  • Practice set: NLP and generative AI mixed questions
  • Timed mini-sim: language and generative AI domain sprint
Chapter quiz

1. A retailer wants to analyze 50,000 customer reviews to identify sentiment and extract key phrases for dashboard reporting. Which Azure service is the best fit?

Show answer
Correct answer: Azure AI Language (Text Analytics features such as sentiment analysis and key phrase extraction)
This is a classic NLP analytics workload (classify sentiment and extract key phrases). Azure AI Language provides built-in text analytics capabilities for sentiment analysis and key phrase extraction. Azure AI Translator is for translation between languages, not sentiment/key phrase analytics. Azure Bot Service is for building conversational experiences, not batch review analytics.

2. A travel company needs to translate chat messages from customers in multiple languages to English in near real time. Which Azure service should you use?

Show answer
Correct answer: Azure AI Translator
The requirement is translation, which maps directly to Azure AI Translator. Azure AI Language focuses on analytics (entities, key phrases, sentiment, classification, question answering) rather than translating text. Azure OpenAI can generate or transform text, but it is not the standard choice for deterministic, scalable translation scenarios expected on AI-900.

3. A company wants to build a customer support chatbot that can handle common questions like order status and store hours and escalate to an agent when needed. Which Azure service is designed for building this conversational solution?

Show answer
Correct answer: Azure Bot Service
This is a conversational/bot workload. Azure Bot Service is the core service for building and hosting bots and integrating with channels. Azure AI Translator only translates text and does not provide dialog orchestration. Azure AI Vision is for image/video analysis, not conversational support.

4. A legal team wants an application that can draft a first version of a contract summary based on provided text and specific instructions. The solution must generate natural language content. Which Azure service is most appropriate?

Show answer
Correct answer: Azure OpenAI Service
Drafting a summary from instructions is a generative AI workload, best matched to Azure OpenAI Service (LLMs for text generation and summarization). Azure AI Translator is for language-to-language translation, not content generation. Azure AI Language sentiment analysis classifies sentiment and does not generate new summaries.

5. A developer is concerned that an LLM-based assistant may produce answers that are not supported by the company’s internal policy documents. Which approach best addresses this concern on Azure?

Show answer
Correct answer: Ground the model by using relevant enterprise data (for example, retrieval-augmented prompts) and apply safety controls
The concern is hallucination/unsupported answers. Grounding (providing authoritative context, commonly via retrieval-augmented generation) helps ensure responses are based on trusted documents, and safety controls align with responsible AI expectations. Increasing temperature generally increases randomness and can worsen factual consistency. Translation may help language understanding but does not ensure answers are grounded in internal documents.

Chapter 6: Full Mock Exam and Final Review

This chapter is where you convert knowledge into score. AI-900 is a fundamentals exam, but it still rewards disciplined execution: time management, pattern recognition across Azure AI services, and clean separation of concepts like training vs. inference, classification vs. regression, vision vs. NLP, and Azure OpenAI vs. “traditional” Cognitive Services. Your goal is not to “learn more” today—it’s to remove score volatility by running two timed mocks, diagnosing mistakes with an error log, and applying a short, high-yield remediation plan.

Across the two simulations (Mock Exam Part 1 and Part 2), you’ll see mixed-domain prompts that test whether you can choose the right workload type and the right Azure service. The final steps—Weak Spot Analysis and the Exam Day Checklist—are what differentiate a near-pass from a comfortable pass. The exam rarely demands deep math, but it does demand precise vocabulary: responsible AI principles, core ML metrics, and service boundaries (what each tool is for, and what it is not for).

Use the sections below as your operating manual. Follow the rules exactly, simulate the pressure, then review like a detective. If you do this well, you’ll walk into exam day with a repeatable process rather than hope.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam rules—timing, no-notes policy, and scoring targets

Section 6.1: Mock exam rules—timing, no-notes policy, and scoring targets

Before you start Mock Exam Part 1, lock in a testing protocol that mirrors the real environment. The point of a mock is not to confirm what you already know; it is to expose what breaks under time pressure. That means: timed, no notes, no searching, no “quick peek” at docs. If you interrupt the simulation, you train your brain to rely on external crutches the real exam won’t allow.

Set a single sitting timer for the full practice block and commit to finishing. If a question stalls you, mark it mentally and move on—don’t spiral. Your scoring target should be aggressive enough to create urgency but realistic enough to track improvement. A practical target for readiness is consistent performance above a comfortable passing buffer, not a single “lucky” high score.

Exam Tip: Track two scores: (1) first-pass accuracy (what you got without revisiting), and (2) final accuracy (after you review marked items). AI-900 punishes slow uncertainty; first-pass accuracy is the best predictor of real performance.

Common trap during timed mocks: over-reading. Many AI-900 prompts include extra context that does not change the decision. Train yourself to identify the decision hook: “which workload type,” “which Azure service,” “which metric,” or “which responsible AI principle.” Once you spot the hook, scan only for constraints (real-time vs batch, structured vs unstructured, need OCR vs object detection, need translation vs sentiment).

Section 6.2: Full-length timed simulation—mixed domains (set A)

Section 6.2: Full-length timed simulation—mixed domains (set A)

Mock Exam Part 1 (Set A) should feel intentionally mixed: you might jump from ML concepts to computer vision services to responsible AI to Azure OpenAI. That switching cost is part of what you’re training. Your job is to classify the question quickly into an objective bucket: AI workloads, ML fundamentals, vision, NLP, or generative AI on Azure.

For ML fundamentals, the exam tests whether you know when to choose classification, regression, clustering, or anomaly detection—and whether you understand training vs. inference. A classic distractor is to describe the training step using inference language (or vice versa). If the prompt mentions “deploy,” “predict,” “score,” or “real-time results,” you’re in inference. If it mentions “labeling,” “fitting,” “learn patterns,” or “evaluate a model,” you’re in training/evaluation.

For metrics, expect high-level interpretation rather than calculation. Precision/recall and confusion-matrix thinking show up as conceptual choices (e.g., prioritizing fewer false negatives vs fewer false positives). Another common trap is mixing accuracy with imbalanced data; accuracy can look high while the model is useless. If the scenario implies rare events (fraud, defects, outages), be suspicious of “accuracy” as the sole success metric.

For service selection across domains, focus on “what is being processed.” Images: consider Azure AI Vision for analysis/OCR; if the scenario implies building custom models, think about Azure AI services that support customization and the broader Azure ML tooling. Text: think language detection, sentiment, key phrases, entity recognition, translation, or conversational patterns. Generative tasks: draft, summarize, or chat → Azure OpenAI patterns, but do not forget safety and content filtering constraints.

Exam Tip: When two options both sound plausible, choose the one that matches the data type and output format in the prompt (bounding boxes vs extracted text vs embeddings vs a predicted numeric value). The output is often the hidden clue.

Section 6.3: Full-length timed simulation—mixed domains (set B)

Section 6.3: Full-length timed simulation—mixed domains (set B)

Mock Exam Part 2 (Set B) is where you verify that your success isn’t tied to memorizing a specific question style. Expect different wording, different service names, and more “best fit” comparisons. Run it fully timed and treat it as a dress rehearsal: a single sitting, controlled environment, and strict pacing.

Set B should also pressure-test your responsible AI instincts. AI-900 regularly checks whether you can identify fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. The trap is choosing a principle that sounds morally nice but doesn’t match the scenario. For example, if the issue is explaining why a model produced a result, that’s transparency/interpretability—not privacy. If the issue is different error rates across groups, that’s fairness—not reliability.

In NLP, watch for confusion between extracting meaning (Text Analytics-style capabilities) and generating language (LLM capabilities). If the prompt wants structured outputs like key phrases, entities, or sentiment labels, it’s analytics. If it wants new text (summaries, rewritten content, conversational responses), it’s generative AI. Similarly, translation is its own workload; don’t label it as sentiment or entity recognition just because it’s “language.”

In vision, differentiate OCR from object detection and from general image analysis. OCR outputs text. Object detection outputs locations/classes. Image analysis may output tags/captions. If a prompt emphasizes “reading labels,” “invoices,” “forms,” or “scanned documents,” your brain should immediately ask: do we need OCR? If the prompt emphasizes “counting items,” “finding defects,” “locating objects,” ask: do we need detection?

Exam Tip: In mixed-domain sets, fatigue causes category errors (vision vs NLP vs generative). If you catch yourself thinking “this is all just AI,” pause and restate the input and required output in one sentence. That single sentence usually reveals the correct domain.

Section 6.4: Review method—error log, distractor analysis, and objective mapping

Section 6.4: Review method—error log, distractor analysis, and objective mapping

Now you do Weak Spot Analysis like an instructor, not like a student. Do not merely note “got it wrong.” Your review must identify the failure mode so it doesn’t recur. Create an error log with these columns: objective domain, concept tested, why your chosen option was tempting, what clue you missed, and the rule you’ll use next time.

Distractor analysis is the fastest way to improve because AI-900 distractors are designed to test boundaries. For each mistake, ask: was I lured by a buzzword (e.g., “AI” → Azure OpenAI), by a near-synonym (“classification” vs “clustering”), or by a service overlap (OCR vs image analysis)? Then write a one-line boundary statement. Example: “If the output must be extracted text, it is OCR; tagging/captioning is not enough.” These boundary statements become your rapid-review notes.

Objective mapping turns mistakes into a plan. Map each error to the course outcomes: AI workloads/responsible AI, ML fundamentals/metrics, vision services, NLP services, generative AI concepts and safety. If one objective owns most of your errors, that’s not a knowledge gap alone—it’s a decision-speed gap. You must practice recognizing that objective in the first 10 seconds of reading.

Exam Tip: Treat every wrong answer as a pattern: “I confuse X with Y when Z is mentioned.” Patterns are fixable; isolated facts are slow to fix. Your error log should read like a list of patterns.

Common review trap: redoing the mock immediately and scoring higher due to memory. Instead, wait at least a day or switch to a different set. Your goal is durable skill, not short-term recall.

Section 6.5: Rapid remediation plan—48-hour weak-spot repair by domain

Section 6.5: Rapid remediation plan—48-hour weak-spot repair by domain

Use the next 48 hours to repair weaknesses with high yield, not broad reading. Divide remediation by domain, using your error log counts to allocate time. A practical split is: the weakest domain gets the largest block, but every domain gets at least a short refresh so you don’t regress elsewhere.

AI workloads & responsible AI: Rehearse the six responsible AI principles and pair each with a scenario cue. Fairness → disparate impact; Reliability & safety → performance under stress; Privacy & security → sensitive data; Inclusiveness → accessibility; Transparency → explainability; Accountability → governance. Practice spotting which principle is being violated vs which is being mitigated.

ML fundamentals: Drill approach selection: classification (labels/categories), regression (numeric prediction), clustering (group without labels), anomaly detection (rare deviations). Reinforce training vs inference language. Revisit metric intuition: accuracy vs precision/recall tradeoffs, and why imbalanced data changes what “good” looks like.

Computer vision: Build a quick decision tree: need text from images → OCR; need locations/bounding boxes → object detection; need description/tags → image analysis. If customization is mentioned, consider the path that allows custom model training (often framed as “custom vision”/custom model approaches).

NLP: Separate analytics (entities, sentiment, key phrases, language detection) from conversational understanding and translation. Your “output format” test is crucial: structured labels vs generated prose.

Generative AI: Review LLM basics (token-based generation, prompts, temperature as creativity control) and core prompt patterns (role, instructions, examples). Include safety basics: content filtering, grounding, and avoiding leakage of sensitive data.

Exam Tip: Your remediation deliverable is not pages of notes—it’s a one-page “decision guide” with if/then rules and boundaries. If you can’t compress it, you don’t own it.

Section 6.6: Final exam tips—pace, question triage, and confidence checklist

Section 6.6: Final exam tips—pace, question triage, and confidence checklist

Your final goal is execution under pressure. Build a pacing plan: a steady rhythm with micro-checkpoints. Avoid spending disproportionate time early; that creates end-of-exam panic where easy points get lost. Use triage: answer the “certain” items immediately, make a best-choice attempt on “likely” items, and flag only the truly “stuck” ones for review.

When you revisit flagged items, do not re-argue from scratch. Apply a rule: restate input → restate required output → pick the service/approach that produces that output. Many candidates change correct answers due to second-guessing; only change an answer when you can name the clue you missed the first time.

Exam Tip: Watch for absolute words in options (“always,” “only,” “guarantees”). Fundamentals exams often use absolutes as distractors because real AI systems have tradeoffs and exceptions.

Confidence checklist for exam day: you can (1) classify a scenario into the correct workload type within 10 seconds, (2) explain training vs inference cleanly, (3) choose a vision capability by the required output (text vs boxes vs tags), (4) choose an NLP capability by the required output (labels vs translation vs conversation), (5) recognize when generative AI is the best fit and name a basic safety control, and (6) match responsible AI principles to scenario cues without hesitation.

Finally, protect your mental bandwidth: read carefully, but don’t over-read; commit to your process; and treat each question as a small identification task, not a research project. The student who manages time and avoids category errors beats the student who “knows a little more” but freezes under the clock.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
Chapter quiz

1. You are reviewing results from a timed AI-900 mock exam. Several missed questions involved predicting house prices from features such as square footage and number of bedrooms. Which machine learning task type best describes this workload?

Show answer
Correct answer: Regression
Predicting a numeric value (house price) is a regression task. Classification predicts discrete labels (for example, 'high/medium/low'), which is not the stated requirement. Clustering is unsupervised grouping without a known target variable, so it does not fit a price-prediction scenario.

2. A company builds a model in Azure Machine Learning and deploys it as a real-time endpoint. During the exam review, a learner confuses the training step with the deployed endpoint. Which statement correctly distinguishes training from inference?

Show answer
Correct answer: Training fits model parameters using labeled data; inference uses the trained model to generate predictions on new data.
Training is the process of learning model parameters from data; inference (scoring) is using the trained model to predict on new inputs. Running a deployed endpoint is inference, not training, so option B reverses concepts and mixes in data preparation tasks. Option C is incorrect because training and inference are distinct stages regardless of compute selection.

3. A support team wants to process incoming emails and automatically extract key phrases and determine whether the sentiment is positive or negative. Which Azure AI service is the best fit?

Show answer
Correct answer: Azure AI Language (Text Analytics capabilities)
Key phrase extraction and sentiment analysis are core NLP capabilities provided by Azure AI Language (formerly Text Analytics). Azure AI Vision is for image analysis, not text. Azure Machine Learning AutoML could be used to build a custom model, but it is not the most direct service for built-in key phrase and sentiment features typically tested on AI-900.

4. You are creating an exam-day checklist. One practice question asks which Responsible AI principle is addressed by providing explanations for why a model produced a specific prediction. Which principle is this?

Show answer
Correct answer: Transparency
Transparency includes interpretability and providing understandable explanations of model behavior. Fairness focuses on avoiding bias and ensuring equitable outcomes across groups, which is different from explanation. Reliability and safety focuses on consistent performance and preventing harm, not on explaining predictions.

5. A marketing team wants to generate new ad copy from prompts and refine outputs using system instructions and examples. They also want to compare this with a 'traditional' service used for tasks like OCR and image tagging. Which option correctly identifies the generative AI service for text generation?

Show answer
Correct answer: Azure OpenAI Service
Azure OpenAI Service is used for generative text tasks (prompt-based content generation, chat, summarization). Azure AI Vision is for image workloads such as OCR and tagging, not generating ad copy. Azure AI Language includes many NLP features, but 'prebuilt entity recognition only' is not a generative text generation service and does not match prompt-driven generation scenarios.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.