AI Certification Exam Prep — Beginner
Learn AI-900 fundamentals plus Copilot and Azure OpenAI—then pass confidently.
This course is a beginner-friendly, exam-prep roadmap for the Microsoft AI-900: Azure AI Fundamentals certification. It is designed for learners with basic IT literacy who want a structured way to understand what Microsoft tests, how to recognize the right Azure AI approach for a scenario, and how to answer common question patterns under time pressure.
The blueprint maps directly to the official AI-900 exam domains:
AI-900 increasingly expects you to recognize where generative AI fits, what it can (and cannot) do, and how Azure services support safe deployment. This course keeps the fundamentals intact while making generative AI practical by explaining core concepts such as tokens, embeddings, grounding, and responsible AI, plus how these ideas appear in Microsoft’s ecosystem through Copilot patterns and Azure OpenAI capabilities.
Chapter 1 orients you to the exam itself—registration, scoring expectations, and a study strategy that aligns your time to the official domains. Chapters 2 through 5 each go deep into one or two domains at a time, using plain-language explanations and exam-style practice prompts that mirror typical AI-900 scenario questions (service selection, workload identification, and basic metric interpretation). Chapter 6 then consolidates everything with a full mock exam experience, analysis of weak areas, and an exam-day checklist.
You do not need prior certification experience. The course focuses on the “why” behind each answer choice so you can reliably eliminate distractors. You’ll also learn the vocabulary Microsoft uses (for example, how prompts describe a workload vs a service, and how responsible AI principles show up in scenario wording).
When you’re ready, start by setting your baseline and planning your schedule. If you’re new to Edu AI, begin here: Register free. Or explore other certification tracks anytime: browse all courses.
By the end, you’ll be able to map a question to an AI-900 domain quickly, select the best-fit Azure AI approach, and complete a timed mock exam with a repeatable review method—helping you walk into the Microsoft AI-900 exam prepared and calm.
Microsoft Certified Trainer (MCT)
Nadia Karim is a Microsoft Certified Trainer who designs exam-first learning paths for Azure and AI fundamentals. She has coached beginners through Microsoft certification prep by translating official objectives into hands-on mental models and exam-style practice.
This chapter sets your operating system for AI-900: how the exam is built, what it rewards, and how to study efficiently—especially if you’re learning Generative AI concepts like Copilot and Azure OpenAI alongside classic AI fundamentals. AI-900 is not a coding exam; it evaluates whether you can recognize AI workloads, select appropriate Azure AI services, and describe core concepts with correct terminology. Most missed questions come from confusing similarly named services, mixing up training vs. inference, or overthinking “best” when the prompt really asks for “most appropriate.”
Your goal is to build a repeatable revision loop: (1) learn the objective, (2) see it in Azure and Microsoft Learn, (3) test recognition with practice items, (4) fix weak spots with targeted notes and short labs. By the end of this chapter you should have a realistic schedule, know how the exam is scored and delivered, and have a practical exam-day toolkit (notes, labs, and checkpoints) that you’ll refine as you progress.
Exam Tip: Treat AI-900 as a vocabulary-and-scenarios exam. If you can map a scenario to a workload (vision/NLP/generative AI) and then to the correct Azure service family, you’ll answer most questions quickly—even when distractors sound plausible.
Practice note for Understand AI-900 format, skills measured, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set up a realistic study plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Learn scoring, passing expectations, and time management: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build your exam-day toolkit: notes, labs, and revision loops: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Baseline diagnostic quiz and goal setting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand AI-900 format, skills measured, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set up a realistic study plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Learn scoring, passing expectations, and time management: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build your exam-day toolkit: notes, labs, and revision loops: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Baseline diagnostic quiz and goal setting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 (Azure AI Fundamentals) measures your ability to describe AI workloads and identify common solution scenarios on Azure. Think of it as “recognize and explain,” not “build and tune.” The exam is organized around official skills domains that typically include: describing AI workloads and considerations, describing fundamental principles of machine learning (ML) on Azure, describing features of computer vision workloads, describing features of natural language processing (NLP) workloads, and describing features of generative AI workloads (including Azure OpenAI and responsible AI concepts). Your course outcomes map directly to these domains: workloads and scenarios; ML fundamentals such as training vs. inference and evaluation; vision and NLP service selection; and generative AI with Azure OpenAI, Copilot concepts, and responsible AI.
What the test is really checking is whether you can: (a) identify the workload type from a short business prompt (for example, “classify images” vs “extract key phrases” vs “generate an email draft”), (b) choose the right Azure AI service category, and (c) describe core ideas accurately (training vs inference, precision/recall basics, prompt vs completion, embeddings vs chat, etc.).
Common traps at this stage are domain-mixing and naming confusion. For example, candidates may treat “Copilot” as a single product rather than a pattern (assistant experience layered over models + grounding + safety + orchestration). Another trap is assuming Azure OpenAI is the answer to every language scenario; the exam often expects you to choose Azure AI Language (Text Analytics capabilities) for extraction/classification tasks, reserving generative AI for content generation, summarization, or conversational experiences that benefit from open-ended output.
Exam Tip: When a question asks you to “describe,” pick the option that uses precise definitions (training builds the model; inference uses it). When it asks you to “choose,” map scenario → workload → service family before reading the answer choices.
Registering correctly is part of exam success because scheduling decisions influence your study plan and stress level. AI-900 is delivered through Pearson VUE. Start at Microsoft Learn’s certification pages to locate the AI-900 exam, confirm the measured skills outline, and follow the “Schedule exam” flow into Pearson VUE. You’ll choose online proctoring or a test center. Online is convenient but has stricter environment rules; test centers reduce the risk of technical interruptions.
Build a realistic timeline by counting backwards from your desired test date. For most learners, a two-to-four-week plan works well if you study 30–60 minutes most days; longer if you’re new to cloud concepts. Align your schedule with the domains: dedicate specific days to ML fundamentals, vision, NLP, and generative AI. Keep one weekly slot for cumulative review, because AI-900 is broad and memory fades quickly without spaced repetition.
Be intentional about time-of-day selection. If you do your best focused work in the morning, schedule the exam then. Also plan a buffer week if possible: it’s not “extra time,” it’s risk management for life events and weak-domain remediation.
Exam Tip: Before scheduling, verify your ID name matches your registration exactly. Mismatches are a surprisingly common reason for day-of delays or forfeits. Also, if you choose online proctoring, run the system test early and again the day before.
AI-900 uses scaled scoring. The key practical point: you need to clear the passing threshold, not achieve perfection. Because it’s scaled, don’t waste energy trying to infer “points per question.” Focus on consistency across domains. If you’re strong in ML and weak in generative AI or vice versa, the exam can expose that imbalance through scenario distribution.
Expect multiple-choice and multiple-response items, plus scenario-based questions. Some items ask you to match capabilities to services (for example, “OCR,” “object detection,” “key phrase extraction,” “sentiment analysis,” “chat completion,” “embeddings”). Others test conceptual boundaries: training vs inference, supervised vs unsupervised learning, classification vs regression, and when to use prebuilt models vs custom models. In the generative AI domain, expect conceptual questions about responsible AI, content filters, grounding, and what Azure OpenAI provides (models, deployments, prompts) versus what Copilot experiences provide (user-facing assistant + orchestration and data access patterns).
Common traps include absolute words (“always,” “never”), confusing similar service names, and selecting an overpowered service. For instance, if a prompt asks to “extract entities and key phrases from text,” the safest choice is Azure AI Language, not Azure OpenAI—unless the question explicitly emphasizes generative output. Another trap is mixing “model” with “deployment.” In Azure OpenAI, you deploy a model and then call that deployment; the exam may test whether you understand that distinction at a high level.
Exam Tip: When two answers both seem correct, pick the one that most directly satisfies the requirement with the simplest service match. AI-900 favors appropriate fit over “coolest” technology.
Your study strategy should mirror the exam’s structure. Organize your notes and revision by the official domains, not by random resources. Create a simple tracker with each outcome and a confidence rating (for example, 1–5). Your goal is to move every item to “4+” before exam day. This is where you integrate a baseline diagnostic (without writing questions in your notes): use a short assessment or practice set to identify which domains are weakest, then allocate time accordingly.
For ML fundamentals, prioritize: what ML is used for, the difference between training and inference, and how evaluation works conceptually (accuracy, precision/recall, confusion matrix basics). Don’t overlearn formulas; learn what each metric indicates and when it matters. For computer vision, know which tasks are classification, object detection, OCR, and image analysis—and which Azure service category supports them. For NLP, distinguish extraction (entities, key phrases), classification (sentiment, language detection), and conversational AI concepts. For generative AI, focus on what makes it different (open-ended output), the Azure OpenAI service role, prompt engineering basics, embeddings/grounding patterns, and responsible AI principles that show up in enterprise deployments.
Plan your weeks in layers: first pass (recognize terms), second pass (scenario mapping), third pass (timed practice and trap avoidance). Insert mini-labs at the end of each domain: even a lightweight “click-through” in the Azure portal or Microsoft Learn sandbox improves recall. Your exam-day toolkit begins here: a one-page “service-to-workload” map, a glossary of confusing pairs (classification vs regression; OCR vs image tagging; Azure OpenAI vs Azure AI Language), and a checklist of responsible AI concepts (fairness, reliability/safety, privacy/security, inclusiveness, transparency, accountability).
Exam Tip: If you can’t explain a concept in one or two sentences (e.g., “inference”), you don’t own it yet. Tight definitions are what the exam rewards.
Microsoft Learn is your primary source of truth because it aligns terminology and service descriptions with what the exam expects. Use it in an objective-driven way: start from the AI-900 skills outline, open the matching Learn modules, and take notes that answer “What is it?”, “When do I use it?”, and “How is it different from nearby options?” Avoid passive reading. After each module, summarize the decision rule in your own words—this is what you’ll recall under time pressure.
Build revision loops. A practical loop looks like: Learn module → 10–15 minutes of recap notes → quick scenario mapping practice (no deep build needed) → correct misunderstandings → update your one-page map. Repeat weekly. The key is to keep notes minimal and usable: tables that map workload to service are better than long paragraphs. For generative AI, add a small “concept map” that distinguishes: model vs deployment, prompt vs completion, chat vs embeddings, and Copilot as an application pattern that can ground model responses in enterprise data with safety controls.
When you hit confusion, go back to the objective and re-anchor. For example, if you’re unsure whether a task is “NLP” or “generative AI,” ask: is the output deterministic extraction/classification, or is it free-form generation? The exam usually separates these intentionally.
Exam Tip: Don’t chase every new product name. AI-900 tests stable fundamentals: workloads, service families, and responsible AI. Use Microsoft Learn to keep your terminology current, but keep your study centered on the measured skills.
Your practice approach should train two skills: accuracy under ambiguity and speed under time limits. Start untimed to learn patterns, then switch to timed sets to build endurance. AI-900 questions often include distractors that are “related but not best.” Your job is to eliminate quickly using workload keywords. For example: “detect objects in an image” points to a vision detection capability; “extract text from images” points to OCR; “summarize a long document” leans generative; “identify key phrases” is classic NLP extraction.
Use a three-pass method during timed practice: (1) answer easy recognition questions immediately; (2) return to medium questions where you can narrow to two; (3) only then spend time on the hardest ones. Manage cognitive load by writing down (on the provided whiteboard tool or scratch paper, where allowed) a short “service shortlist” for the scenario before reading all options. This prevents being pulled toward shiny distractors.
Time management is less about rushing and more about avoiding unproductive loops. If you’re stuck, mark it and move on. When reviewing, don’t just note the right answer—note the rule that made the other option wrong. This is how you build trap resistance.
Exam Tip: For multiple-response items, treat each option as true/false against the requirement. Many candidates fail by selecting one extra plausible statement. Only select choices that are explicitly supported by the scenario and the service’s core purpose.
1. You are creating a study plan for AI-900. A colleague asks whether they need to write code during the exam. Which statement best describes what AI-900 evaluates?
2. During practice questions, you frequently miss items where Azure services have similar names. Which action most directly targets the likely cause of these misses for AI-900?
3. A team member tends to overthink questions that ask for the "best" choice and often changes correct answers to incorrect ones. What exam strategy is most appropriate for AI-900-style questions?
4. You are mentoring a candidate who confuses training with inference in multiple practice questions. Which statement correctly distinguishes them in an exam scenario context?
5. You want a repeatable revision loop for AI-900 that reduces weak areas quickly while keeping effort realistic. Which sequence best matches an effective approach for this exam?
This chapter is where many AI-900 candidates either gain easy points—or lose them due to vague wording. The exam does not expect you to build models, but it does expect you to recognize workloads, match them to common solution scenarios, and pick the Azure service category that fits. In Copilot and Azure OpenAI questions, you’ll also be tested on whether you can tell “generative AI” apart from traditional ML, and whether you can apply Responsible AI principles to realistic business situations.
Throughout this chapter, practice a simple exam habit: when you read a scenario, underline the input type (text, image, audio, tabular data), the desired output (label, number, summary, generated content), and whether there is an explicit “right answer” label available (supervised vs unsupervised). Those three clues usually identify the workload in under 10 seconds.
Exam Tip: Many wrong answers are “adjacent” services or adjacent workloads. If a scenario asks for extracting insights from documents, don’t jump straight to “chatbot.” If it asks for generating new text, don’t pick “text analytics” (which analyzes existing text). Train yourself to spot the verb: classify, predict, cluster, detect, extract, search, converse, generate.
Now, let’s do the domain deep dive section by section.
Practice note for Differentiate AI, ML, deep learning, and generative AI in exam terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map business problems to AI workloads and Azure solution types: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: AI workloads scenario questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-review: domain summary and memory cues: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate AI, ML, deep learning, and generative AI in exam terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map business problems to AI workloads and Azure solution types: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: AI workloads scenario questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
On AI-900, “AI workload” is shorthand for the type of problem you are solving with AI. The exam commonly frames workloads as: machine learning (prediction from data), computer vision (images/video), natural language processing (text/language), knowledge mining (search + extraction from unstructured content), conversational AI (bots), and generative AI (creating new content). You’re expected to connect a business goal to one of these workload categories and then to an Azure solution type.
Use exam definitions (not research-paper definitions). AI is the broad umbrella of systems that exhibit “intelligent” behavior. ML is a subset where a model learns patterns from data. Deep learning is a subset of ML that uses neural networks with multiple layers and is often used for vision, language, and generative models. Generative AI is a subset focused on producing new outputs (text, images, code) based on prompts. In exam terms: if the scenario says “generate,” “draft,” “summarize,” “rewrite,” “create an image,” or “produce code,” you are likely in generative AI territory.
Exam Tip: Don’t overthink the math. The exam is checking whether you can map: (1) business problem, (2) workload, (3) Azure capability. If a scenario mentions “train a model using historical labeled data,” that’s classic ML. If it mentions “use a prebuilt API to extract key phrases,” that’s an Azure AI service. If it mentions “ground the model in company documents,” think retrieval/knowledge mining patterns plus generative AI.
Common trap: candidates answer with a specific product name when the question is asking for the workload type. Read the prompt carefully: if it says “which AI workload,” answer with the workload (e.g., NLP), not the service (e.g., Azure AI Language). If it says “which service,” then choose the service category that matches.
This is a high-frequency AI-900 objective because it tests whether you can interpret outputs. The fastest way to decide is to look at the shape of the output and whether there is labeled training data. Classification outputs a category (often yes/no or a set of classes). Regression outputs a continuous numeric value. Clustering outputs groups based on similarity without predefined labels. Anomaly detection outputs a flag/score indicating “unusualness,” often in time series or transactional data.
Training vs inference is also tested here. Training is when you fit the model using historical data (often compute-heavy, iterative). Inference is when you use the trained model to make predictions on new data (often latency-sensitive). The exam likes to embed this distinction in operational scenarios: “batch scoring overnight” vs “real-time scoring at checkout.”
Exam Tip: If the scenario includes “known outcomes” like churned/not churned or fraud/not fraud, that strongly suggests supervised learning (classification). If it includes “we don’t know the groups ahead of time,” that suggests clustering. If it includes “detect spikes,” “outliers,” “unusual patterns,” that points to anomaly detection—even if the output is just a score.
Model evaluation appears often as “which metric is appropriate?” At AI-900 level, remember: classification commonly uses accuracy/precision/recall; regression commonly uses error metrics (like MAE/RMSE); anomaly detection often uses precision/recall style thinking because anomalies are rare. The trap is picking “accuracy” for imbalanced fraud scenarios: accuracy can be misleading if 99.9% transactions are normal.
Exam Tip: When you see “rare event” language (fraud, intrusion, defect), think class imbalance and avoid simplistic “accuracy” reasoning. The exam isn’t asking for metric formulas, but it is checking whether you can spot what “good” means in context.
These workloads are easy points if you lock onto the input modality. Computer vision uses images or video frames: object detection, image classification, face detection (not identification unless allowed), OCR (extracting printed/handwritten text), and spatial understanding. NLP uses language as input: detecting sentiment, extracting entities, identifying key phrases, summarizing existing text (note: summarization can be NLP or generative depending on phrasing and tool choice). Knowledge mining is about search and extraction from large collections of unstructured content—documents, PDFs, scans—typically involving indexing plus enrichment. Conversational AI is about dialogue: managing multi-turn interactions, intents, and responses.
The exam often tests whether you can separate “analyze” from “generate.” For example, extracting entities from customer emails is NLP/text analytics. Generating a reply email is generative AI. Indexing 100,000 policy PDFs so employees can search them is knowledge mining (often paired with Azure AI Search). Creating a chat experience that answers HR policy questions could be conversational AI, but if it also “reads” policy PDFs to answer accurately, the underlying pattern is knowledge mining + conversational interface (and possibly generative AI for natural responses).
Exam Tip: “Search across documents” is a keyword for knowledge mining. “Chatbot” is not automatically the right answer if the real need is document indexing and retrieval. Many scenarios describe a bot interface, but the core workload is still search + extraction.
Common trap: OCR belongs to computer vision even though the output is text. If the scenario is “extract text from images,” the input is still an image workload. Another trap: “speech” scenarios can blur boundaries. If the exam mentions converting speech to text, that’s speech AI; once text is produced, NLP can analyze it. Read for the primary task.
Generative AI questions typically follow a “prompt-in, content-out” pattern: a user provides instructions (and possibly context), and the system outputs newly composed text, code, images, or structured content. For AI-900 in the Copilot/Azure OpenAI domain, you must recognize when the requirement is creation rather than analysis. Drafting marketing copy, writing a job description, summarizing a meeting into action items, generating SQL from natural language, and rephrasing content in a brand voice are all generative patterns.
Azure OpenAI is the canonical “service family” the exam expects you to associate with generative text/code/image scenarios. Copilot scenarios often describe the same idea but embedded into productivity tools: “help users write,” “help users summarize,” “help users generate a plan.” Treat Copilot as an application experience built on generative models plus grounding, security, and orchestration. The exam generally won’t require deep architecture, but it will check that you understand the role: Copilot helps users produce outcomes inside workflows.
Exam Tip: If the scenario mentions “use company data to answer accurately,” look for retrieval/grounding language (documents, policies, product catalog). This is a common pattern: retrieval-augmented generation (RAG). The trap is assuming the model “already knows” internal documents. On the exam, when proprietary data is required, the correct approach usually involves retrieving relevant content and then generating an answer grounded in that content.
Common trap: “Summarize this text” can be answered by either extractive summarization (traditional NLP) or generative summarization (Azure OpenAI). If the question emphasizes “generate a well-written summary,” “rewrite,” or “create,” prefer generative. If it emphasizes “analyze text to extract key phrases,” prefer NLP/text analytics.
Responsible AI is not a side topic on AI-900; it is a scoring area that appears in both classic ML and generative AI questions. You must be able to match a scenario risk to a principle. Memorize the six principles and tie each to a “what could go wrong” example.
Exam Tip: The exam often disguises these principles as operational requirements. “We need to explain why a decision was made” maps to transparency. “We need a human to review before sending” maps to accountability and safety. “We must not store customer SSNs in logs” maps to privacy/security. Choose the principle that is most directly threatened in the scenario.
For generative AI, reliability and transparency show up as “hallucinations” and “over-trust.” A safe design might include grounded answers, citations to retrieved sources, user warnings, and fallback behavior when confidence is low. For ML, fairness and accountability show up in decision systems (hiring, credit, insurance). The common trap is selecting “privacy” for every sensitive scenario; instead, identify whether the issue is data handling (privacy) or outcome bias (fairness) or inability to explain/trace decisions (transparency/accountability).
This section is your practical “scenario matching” toolkit. You are not answering practice questions here; you are building a repeatable method that works under time pressure. AI-900 scenarios are short, so your goal is to extract the signal quickly and avoid distractors.
Step 1: Identify the artifact. Are you dealing with tabular data (rows/columns), text, images, documents, or a conversation? Tabular usually points to ML (classification/regression/clustering/anomaly). Images point to computer vision. Large document sets point to knowledge mining and search. Multi-turn interaction points to conversational AI.
Step 2: Identify the verb. Predict/forecast/estimate → regression. Decide/label/categorize → classification. Group/similar segments → clustering. Unusual/outlier/spike → anomaly detection. Extract entities/key phrases/sentiment → NLP. Detect objects/recognize text in image → vision. Search/index/enrich documents → knowledge mining. Draft/rewrite/summarize/generate → generative AI.
Step 3: Decide if it’s “analyze” or “create.” This is where Azure OpenAI and Copilot become obvious. If the system is producing original prose/code/images from a prompt, it’s generative. If it’s identifying properties of existing content, it’s analytics.
Exam Tip: Watch for mixed scenarios. Many real solutions combine workloads (e.g., OCR + NLP; search + generative answers; speech-to-text + sentiment). The exam usually asks for the primary workload or the best first step. Don’t pick a later-stage component if the scenario is explicitly about the earlier transformation (e.g., “convert scanned forms to text” is vision/OCR, even if the next step is NLP).
Common traps to avoid: (1) picking “deep learning” as a workload when the question wants the business task (vision/NLP/etc.); (2) confusing chatbot UX with the underlying knowledge mining requirement; (3) calling all language tasks “generative” when many are simple text analytics; and (4) forgetting Responsible AI in scenarios with hiring, lending, healthcare, or any user-facing generated content. If you can label the workload, you can usually eliminate two options immediately—and that’s how you turn domain knowledge into exam points.
1. A retail company wants to automatically route incoming customer emails into categories such as "billing", "returns", and "technical support". They have thousands of previously labeled emails for training. Which AI workload best fits this requirement?
2. A manufacturer wants to predict next month’s energy consumption for each facility using historical meter readings and weather data. The output should be a numeric value (kWh). Which AI workload should you choose?
3. A legal team has thousands of contracts (PDFs) and wants to extract key fields such as party names, effective date, and termination clause text to populate a database. Which Azure solution type best aligns with this workload?
4. A marketing team wants an assistant that can draft new product descriptions in the company’s style based on a short bullet list of features. Which approach is the best match?
5. A bank deploys an AI system to assist with loan pre-screening. After a pilot, the bank discovers that applicants from a specific neighborhood are disproportionately rejected due to historical data patterns. Which Responsible AI principle should be prioritized to address this issue?
This chapter targets the AI-900 expectation that you can describe machine learning (ML) at a fundamentals level and connect those fundamentals to Azure services and common exam scenarios. Even though this course focuses on Copilot and Azure OpenAI, AI-900 still tests classic ML vocabulary and lifecycle concepts because they underpin how data-driven systems (including generative AI solutions) are built, evaluated, and deployed.
As you read, keep an exam mindset: AI-900 questions often provide a short scenario and then ask you to identify (1) the ML approach (supervised vs unsupervised), (2) the correct evaluation metric (classification vs regression), or (3) the correct Azure capability (training vs inference, workspace vs compute). The most common mistakes come from missing a single word cue in the prompt—terms like “predict a category,” “forecast a value,” “group similar items,” “deploy,” or “real-time scoring.”
We’ll first lock in core ML definitions (features, labels, training, inference), then connect them to approach selection and evaluation metrics, then map to Azure Machine Learning fundamentals (workspace, compute, pipelines, Automated ML, and Designer). You’ll finish with a practice-oriented mini-review of word cues and decision patterns you can apply under exam time pressure.
Practice note for Understand core ML concepts: features, labels, training, inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose ML approaches and metrics for typical AI-900 prompts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain Azure Machine Learning capabilities at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML lifecycle and metrics questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-review: key definitions and exam word cues: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand core ML concepts: features, labels, training, inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose ML approaches and metrics for typical AI-900 prompts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain Azure Machine Learning capabilities at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML lifecycle and metrics questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-review: key definitions and exam word cues: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to use ML terminology precisely. A dataset is a collection of examples (rows) with attributes (columns). In supervised learning, each example has features (inputs) and a label (the output you want to predict). Features might be “age,” “income,” and “tenure”; a label might be “will churn” or “monthly sales.” If the prompt says “known outcomes,” “historical results,” or “labeled data,” that’s your cue that labels exist and supervised learning is likely.
Training is the process of fitting a model to data—learning patterns that map features to labels. Inference (also called scoring or prediction) is using the trained model to make predictions on new data. On the exam, these are commonly confused with deployment. Deployment is about packaging and hosting a model so inference can happen in an application; training is not deployment.
Exam Tip: When a scenario includes words like “build,” “train,” “fit,” “learn,” or “create a model,” think training. When you see “predict,” “classify,” “score,” “real-time,” “batch scoring,” or “during app usage,” think inference.
Finally, remember the broad ML lifecycle: define the problem, prepare data, train, evaluate, and deploy for inference. Questions may ask you to order steps; choose the option where evaluation occurs after training but before deployment.
Approach selection is a frequent AI-900 prompt pattern. Supervised learning uses labeled data to learn a mapping from features to a known target. It covers two big families: classification (predict a category) and regression (predict a number). If the question says “approve/deny,” “spam/not spam,” “diagnosis type,” or “which product category,” you are in classification territory. If it says “forecast demand,” “predict price,” “estimate time,” or “temperature,” that points to regression.
Unsupervised learning uses unlabeled data to find structure. The most tested unsupervised technique in fundamentals exams is clustering—grouping similar items. Word cues include “segment customers,” “group documents,” “find similar behaviors,” or “identify natural groupings.” Another unsupervised idea you might see is anomaly detection (finding unusual patterns), which can be framed as “find outliers” or “detect unusual transactions.”
Exam Tip: Look for the “what is the output?” clue. If the output is a known label from history, it’s supervised. If the output is a grouping or “discover patterns,” it’s unsupervised.
In Azure-centric wording, you may see “train a model to predict” (supervised) versus “analyze data to find segments” (unsupervised). Your job is to match the approach to the scenario, not to name a specific algorithm.
Evaluation metrics are chosen based on the task type and the cost of mistakes. For classification, accuracy (correct predictions / total predictions) is a basic metric, but it can be misleading when classes are imbalanced (for example, fraud detection where “fraud” is rare). In those cases, the exam often wants you to prefer precision and recall. Precision answers: “When the model predicts positive, how often is it correct?” Recall answers: “Of all actual positives, how many did we catch?”
Exam Tip: If false positives are expensive (flagging many legitimate transactions as fraud), precision matters. If false negatives are expensive (missing a fraudulent transaction or missing a disease), recall matters. Prompts sometimes describe the business impact; map that impact to the error type.
For regression, a common metric is RMSE (root mean squared error), which penalizes larger errors more than smaller ones. If the prompt says “predict a numeric value” and asks which metric to use, RMSE is a safe exam choice. You may also see MAE, but RMSE is the frequent fundamentals cue.
AI-900 also expects you to recognize overfitting: a model performs very well on training data but poorly on unseen data. The scenario clue is typically “high training accuracy but low test accuracy.” Overfitting is not solved by “train longer” by default; it’s often addressed by more data, simplifying the model, regularization, or better validation practices.
When the exam asks how to improve generalization, the best fundamentals answer usually references addressing overfitting and validating on separate data, not tweaking a deployment setting.
At the fundamentals level, Azure Machine Learning (Azure ML) is the Azure service used to build, train, and manage ML solutions end-to-end. AI-900 won’t test deep configuration, but it will test what major components are for and how they relate to training and deployment.
A workspace is the top-level container for Azure ML resources: experiments, models, datasets, registered assets, and related configuration. If a prompt says “central place to manage models, runs, and artifacts,” the workspace is the answer. Compute refers to the resources used to run training or inference workloads—examples include compute instances (development) and compute clusters (scalable training). If a question asks what you scale to reduce training time, choose compute (often a cluster), not the workspace.
Pipelines are used to define and automate ML workflows as repeatable steps (data preparation, training, evaluation, and sometimes deployment). Pipelines help with reproducibility and operationalizing the ML lifecycle. On the exam, the cue is “repeatable workflow,” “orchestrate steps,” or “automate training process.”
Exam Tip: Map nouns to responsibilities: workspace = management boundary, compute = where it runs, pipeline = how steps are chained. This quick mapping helps you answer many “which component” questions without needing product depth.
In a Copilot/Azure OpenAI world, Azure ML still matters for classic predictive models and for managing ML operations. The exam wants you to recognize where Azure ML fits when the solution is not purely generative—especially for forecasting, classification, and tabular ML scenarios.
Two Azure ML capabilities frequently tested at a high level are Automated ML (AutoML) and Designer. AutoML helps you automatically try multiple algorithms and hyperparameters to find a strong model for your dataset and task type (classification, regression, forecasting). The key exam idea: AutoML is for accelerating model selection and tuning when you know the target (label) and want a good baseline quickly.
Designer is a drag-and-drop visual interface for building ML workflows without writing much code. It’s about assembling data prep, training, and evaluation steps visually. If the prompt says “no-code,” “visual pipeline,” or “drag-and-drop,” Designer is the cue. If it says “automatically choose the best algorithm” or “optimize model,” AutoML is the cue.
Exam Tip: AutoML answers “Which model configuration is best?” Designer answers “How do I build the workflow visually?” Don’t swap them.
Also remember that AutoML and Designer relate primarily to the training phase. The exam may include “deploy the best model” as a follow-on step, but AutoML/Designer are not synonymous with deployment; they help create and evaluate models that can later be deployed for inference.
This section consolidates the chapter into the patterns AI-900 repeatedly uses: choose the metric, choose the lifecycle step, and choose the ML approach. When you see a prompt, first classify it as classification/regression/clustering, then select the metric and Azure capability that match that type. The exam is less about math and more about correct matching.
Metrics selection cues: If the output is a category, think classification metrics. Accuracy is acceptable for balanced classes; precision/recall are better when the scenario describes imbalance or costly mistakes. If the output is a number, think regression metrics like RMSE. If the prompt is about grouping, metrics might not be central; instead, the right approach is clustering and the evaluation discussion is typically more qualitative at fundamentals level.
Lifecycle ordering cues: Most correct sequences include data preparation before training, training before evaluation, and evaluation before deployment. If an option evaluates before training, it’s wrong. If an option deploys before any evaluation, it’s usually wrong in an exam context unless the question explicitly says “quick proof of concept” (even then, evaluation is typically implied).
Azure component cues: “Track experiments, register models, manage artifacts” points to the workspace. “Run training at scale” points to compute clusters. “Automate a repeatable workflow” points to pipelines. “Automatically try algorithms/hyperparameters” points to AutoML. “No-code visual workflow” points to Designer.
Exam Tip: Under time pressure, underline the output noun in the scenario (category/number/group) and the operational verb (train/evaluate/deploy/predict). Those two highlights typically determine the correct answer set.
In your mini-review, memorize the word cues: “segment” → clustering; “forecast” → regression (RMSE); “classify”/“approve” → classification (accuracy/precision/recall); “high train, low test” → overfitting; “central management” → workspace; “scale training” → compute; “orchestrate steps” → pipeline; “best algorithm automatically” → AutoML; “drag-and-drop” → Designer.
1. A retail company wants to predict whether an online transaction is fraudulent (Yes/No) based on features such as purchase amount, country, device type, and account age. Which machine learning approach should you choose?
2. A utility provider wants to forecast next month’s electricity consumption (kWh) for each household using past monthly usage, home size, and number of occupants. Which metric is most appropriate to evaluate the model?
3. A marketing team has a dataset of customer purchase behavior but no labels. They want to group customers into segments with similar behavior patterns to target promotions. Which approach best fits this requirement?
4. You are preparing an ML solution in Azure. You want a central place to manage datasets, experiments, model training runs, and registered models for a project. Which Azure Machine Learning component should you use?
5. A company has trained a model and now needs to use it in production to return predictions for new data through a REST call. In ML terminology, which phase is this?
In AI-900, “computer vision” means turning images and video into structured information that applications can act on. The exam does not test you on deep math or model architectures; it tests whether you can recognize a vision task, choose the right Azure AI service, and explain outputs at a high level. This chapter ties directly to the objective “Describe computer vision workloads on Azure and choose appropriate Azure AI services,” and it also supports the broader course outcomes around training vs. inference and service selection.
As you work through the lessons in this chapter, keep an exam mindset: the wording of the scenario almost always signals the workload. Words like “identify the main subject” or “categorize images” typically point to classification; “find and locate items” points to detection; “extract text” points to OCR; and “find per-pixel boundaries” points to segmentation. Exam Tip: On AI-900, you’re often graded on choosing the most appropriate managed service (Azure AI Vision, Azure AI Face, Azure AI Custom Vision, Azure AI Document Intelligence) rather than describing how to build a model from scratch.
You’ll also see a recurring theme: out-of-the-box (prebuilt) vs customized (trained) solutions. Many scenarios can be handled with prebuilt models using Azure AI services; customization becomes relevant when labels are domain-specific (e.g., defect types in manufacturing) or when you need to train on your own images. The lessons in this chapter guide you through identifying tasks, explaining detection/OCR/classification plainly, understanding customization and deployment at a high level, practicing service selection, and finishing with a quick vision cheat sheet.
Practice note for Identify vision tasks and pick the right Azure AI service: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain detection, OCR, and image classification in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand vision customization and model deployment at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: computer vision service selection questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-review: vision cheat sheet and last-minute recalls: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify vision tasks and pick the right Azure AI service: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain detection, OCR, and image classification in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand vision customization and model deployment at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
The AI-900 objective expects you to recognize common vision workloads and map them to Azure services. At a foundational level, computer vision workloads include: describing what’s in an image (tags/captions), detecting objects and their locations (bounding boxes), reading text (OCR), identifying people-related attributes (face detection/verification), and analyzing documents/receipts/forms (document understanding).
For the exam, focus on “workload recognition” more than implementation details. You should be able to explain what the system outputs during inference: labels (classification), coordinates + labels (detection), pixel masks (segmentation), extracted text (OCR), or structured fields (document extraction). Exam Tip: When a prompt mentions “real-time camera feed,” “safety monitoring,” or “counting items,” it’s typically an inference scenario where you call an API endpoint—no training required unless the classes are unique to your business.
Also know the rough boundary between “image understanding” and “document understanding.” Images like photos of products, streets, or animals are usually handled by Azure AI Vision (or Custom Vision if you need custom labels). Documents like invoices, receipts, IDs, and forms often fit Azure AI Document Intelligence because the goal is structured field extraction rather than just text lines.
A common trap is overcomplicating the choice: AI-900 typically rewards selecting the managed AI service aligned to the workload rather than proposing Azure Machine Learning training pipelines unless the scenario explicitly asks to train and manage models end-to-end.
Classification, detection, and segmentation are easy to confuse under exam pressure because all three “recognize things in images.” The difference is the granularity of the answer. Classification returns “what it is” for the whole image (or a region you provide). Detection returns “what and where” using bounding boxes. Segmentation returns “what and exactly which pixels,” producing a mask for each object/region.
Use scenario keywords to decide quickly. If the question wants “categorize images in a folder” or “determine if an image contains a forklift,” classification is the match. If it wants “locate every forklift and count them,” detection is the match. If it wants “measure the area of corrosion” or “separate background from foreground precisely,” segmentation is the match.
Exam Tip: Bounding boxes imply detection. “Outline,” “mask,” “pixel-level,” or “exact shape” implies segmentation. If neither location nor shape is required, default to classification.
At a service level, Azure AI Vision provides general-purpose analysis and can perform tasks like detecting objects and generating tags/captions. Azure AI Custom Vision is commonly associated with training your own classification or object detection model when the built-in labels are not sufficient (for example, custom defect categories). A frequent exam trap is selecting Custom Vision when the scenario does not require custom labels; if a prebuilt capability can solve it, AI-900 often expects the prebuilt service selection.
Finally, remember the high-level training vs inference distinction: classification/detection/segmentation can be prebuilt (no training by you) or custom (you label images, train, then deploy). On AI-900, you mainly need to articulate that training happens once (or periodically), while inference is the repeated act of scoring new images.
OCR (optical character recognition) turns text in an image into machine-readable text. AI-900 often frames OCR as “read text from photos,” “extract serial numbers,” “read street signs,” or “digitize scanned forms.” The key exam skill is distinguishing plain OCR from document understanding. OCR returns text lines/words and their positions; document understanding returns structured fields like totals, dates, and line items.
Azure AI Vision can perform OCR on images (including printed text in photos). When the scenario involves business documents—receipts, invoices, purchase orders, ID documents—Azure AI Document Intelligence is typically the better fit because it goes beyond raw text to extract structure, key-value pairs, and tables. Exam Tip: If the prompt mentions “fields,” “table extraction,” “key-value pairs,” “invoice total,” or “receipt merchant name,” that’s a Document Intelligence signal, not just OCR.
Another common trap is assuming OCR is the same as search. OCR is a vision workload (extract text), while search solutions require indexing and querying; the exam typically wants you to pick the vision service first, then optionally you could store results elsewhere for retrieval. Don’t jump straight to a search service if the core requirement is “read text.”
From an “interpret outputs” standpoint, OCR results commonly include the recognized text and positional data (bounding boxes for words/lines). The exam may describe output as “coordinates” or “regions” for text; that is still OCR. Document extraction outputs are more semantic: “InvoiceTotal: 123.45,” “VendorName: …,” often with confidence per field.
Face-related scenarios appear in AI-900 as “detect faces in an image,” “compare two faces to verify the same person,” or “group similar faces.” On Azure, these are typically associated with Azure AI Face (a dedicated service for face detection and analysis). At a fundamental level, face detection finds face regions; face verification compares whether two faces likely belong to the same person; identification matches a face against a known set (when allowed and configured).
The exam also expects responsible AI awareness. Face capabilities can be sensitive and may have usage constraints, approval requirements, or limitations depending on region and policy. You should know that responsible use includes consent, privacy, bias considerations, and using the technology only for appropriate scenarios. Exam Tip: If a question hints at surveillance, tracking individuals without consent, or sensitive attribute inference, the “best answer” often involves emphasizing responsible use, governance, and selecting compliant capabilities rather than blindly enabling face features.
A common trap is confusing “face detection” with “object detection.” Face detection is a specialized capability: it focuses on human faces and can return face rectangles and face-related metadata. Object detection is broader (cars, helmets, animals). If the scenario explicitly mentions “faces” or “verify identity,” choose Azure AI Face rather than generic vision object detection.
Keep your explanation at the AI-900 level: you are not expected to describe embeddings in detail, but you should be able to say the service produces a numeric representation internally to compare similarity scores, and outputs a match/confidence-type result.
This section connects the “pick the right service” lesson with a repeatable decision pattern. On AI-900, the most common vision-related services you’ll see are Azure AI Vision (general image analysis, tagging, captions, OCR, object detection), Azure AI Custom Vision (train custom classifiers or detectors using your own labeled images), Azure AI Face (face-specific), and Azure AI Document Intelligence (structured document extraction). Your job is to map the scenario to the simplest service that meets requirements.
Start with the question: is it a photo-like image or a business document? If it’s a document with fields/tables, lean Document Intelligence. If it’s a photo and the task is “describe what’s in it,” lean Azure AI Vision. If it requires custom labels (like proprietary parts or niche defects), Custom Vision becomes the likely answer because you can upload images, label them, train, and publish an endpoint.
Exam Tip: The phrase “custom model” or “train with our own images” is a direct pointer to Custom Vision. If the requirement is generic (animals, common objects, captions), prebuilt Vision is usually correct.
At a high level, customization and deployment follow a predictable loop: collect representative images, label them, train the model, evaluate accuracy, then publish for inference. AI-900 expects you to recognize that training and evaluation occur before deployment, and that deployment exposes an endpoint used by apps. A trap is mixing up training location: Custom Vision is a managed training experience; Azure Machine Learning is the broader platform for building and managing custom ML, but it’s not the default answer for a straightforward “custom image classifier” scenario.
When two services seem plausible, choose the one that best matches the output format requested. “Extract fields” beats “extract text.” “Train on our data” beats “use prebuilt.” “Verify person” beats “detect object.” This is how you avoid 50/50 errors on the exam.
To prepare for the practice set in this chapter, train yourself to do three actions quickly: (1) identify the vision task, (2) select the best-fit Azure AI service, and (3) interpret what the output would look like. AI-900 questions often hide the answer in nouns: “caption,” “tags,” “bounding box,” “mask,” “key-value,” “table,” “verify,” “gallery,” and “OCR.” Your goal is to match those nouns to the right workload and service.
When interpreting outputs, remember the exam’s level of detail: you should recognize that classification returns labels + confidence; detection returns labels + confidence + coordinates; OCR returns text + layout coordinates; document extraction returns fields/tables; and face verification returns a similarity/match result. Exam Tip: If a scenario says “we need coordinates,” ask: coordinates of what? Objects implies detection; words/lines implies OCR; invoice fields implies Document Intelligence.
Common scenario-matching traps include: choosing Custom Vision when the scenario does not mention custom labels; choosing OCR when the scenario really needs structured invoice fields; and choosing generic Vision for identity verification instead of Face. Another trap is ignoring the deployment hint. If the question mentions “publish an endpoint” after training, that’s a Custom Vision workflow; if it mentions “call a prebuilt API,” that’s typically Azure AI Vision or Document Intelligence prebuilt models.
Finish your mini-review with a “vision cheat sheet” mindset: identify the keywords, map to task, map to service, and predict the output. If you can do that consistently, you’ll be ready for AI-900 computer vision questions even when the scenario is wrapped in business context (retail, manufacturing, healthcare, or security).
1. A retail company wants to process in-store camera images to locate and count products on shelves. The output must include bounding boxes for each detected product. Which computer vision workload best matches this requirement?
2. A logistics company scans shipping labels and needs to extract the recipient name, street address, and tracking number from the label image. Which Azure AI service is the most appropriate choice for this scenario?
3. A wildlife organization has thousands of images from trail cameras and wants to automatically categorize each image into one of several animal species (for example, deer, bear, fox). The organization does not need object locations, only the category per image. Which vision task is being described?
4. A manufacturer wants to identify specific types of defects on circuit boards that are unique to their production process. Off-the-shelf labels like "damaged" are not sufficient, and they want to train using their own labeled images and deploy the resulting model for predictions. Which Azure service best fits this requirement?
5. You are designing a solution and must choose the most appropriate managed Azure AI service. A customer wants to take photos of store signs and extract the visible text so it can be translated and displayed to users. Which service should you recommend?
This chapter maps directly to the AI-900 outcomes around NLP workloads and generative AI workloads, with special attention to the exam’s “choose the right service” style questions. Expect the test to probe whether you can distinguish classic NLP (extract, classify, summarize) from GenAI (generate, transform, chat), and whether you understand the deployment and safety vocabulary used with Azure OpenAI and Copilots. A common exam pattern is a short scenario (customer support emails, internal policy search, meeting notes) followed by “which Azure service fits best?” Your job is to recognize the workload type, the required output (labels vs free-form text), and the risk constraints (privacy, hallucinations, safety).
As you read, keep a mental decision tree: (1) Is the task deterministic extraction/analysis? Think Azure AI Language features (Text Analytics). (2) Is it conversation/knowledge base? Think bot concepts and Q&A patterns. (3) Is it generation, rewriting, or code/content creation? Think Azure OpenAI. (4) Does the business want GenAI embedded into everyday tools (M365, Power Platform, Dynamics)? Think Copilot and governance.
Practice note for Understand NLP tasks and Azure service choices for text and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts: tokens, embeddings, grounding, retrieval: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Use Azure OpenAI and Copilot concepts in exam language (no coding required): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: NLP + GenAI + responsible AI questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-review: final domain map and cross-domain comparisons: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand NLP tasks and Azure service choices for text and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts: tokens, embeddings, grounding, retrieval: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Use Azure OpenAI and Copilot concepts in exam language (no coding required): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: NLP + GenAI + responsible AI questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-review: final domain map and cross-domain comparisons: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 frequently tests “NLP workloads” as analysis of text rather than generation. In Azure, these are commonly delivered through Azure AI Language capabilities (often referred to in exam prep as “Text Analytics” features). The scenario clues are words like detect, classify, extract, score, or identify—outputs that are structured (labels, entities, key phrases) rather than open-ended prose.
Core workload types you should recognize: sentiment analysis (positive/neutral/negative scoring), key phrase extraction (main topics from text), entity recognition (people, organizations, locations, dates, product names), and language detection. You may also see “PII detection” as a compliance-oriented extraction task (find sensitive data in documents). These fit classic NLP because they turn unstructured text into structured insights that you can store, filter, and trend over time.
Exam Tip: If the question asks for “insights from text” and expects a list of entities, key phrases, or sentiment scores, choose Azure AI Language/Text Analytics-type options—not Azure OpenAI. GenAI can do these tasks, but the exam expects you to pick the purpose-built analytics service when the output is structured and repeatable.
Common trap: confusing OCR and text analytics. OCR (extracting text from images) is a computer vision/document workload, while sentiment/entity extraction is NLP. Another trap is assuming “summarization” always means GenAI. Azure AI Language can provide summarization features; the exam will often reward you for selecting an NLP summarization capability when the requirement is straightforward summarizing rather than creative generation.
How to identify the correct service choice: look for the input type (plain text), the output type (scores/fields), and the operational need (batch processing, dashboards, compliance). If the scenario mentions reporting, trends, or KPI monitoring from customer feedback, sentiment + key phrases + entities is usually the right mental bundle.
Conversational AI on AI-900 is less about building bots and more about recognizing the building blocks. A “bot” is the interaction layer (channels like web chat, Teams, etc.). Underneath, the system needs to understand what the user wants (intent) and pull out important details (entities). For example, in “Book a flight to Seattle tomorrow,” the intent might be BookFlight and the entities include destination=Seattle and date=tomorrow.
Many exam questions focus on Q&A patterns (sometimes described as knowledge base or FAQ style). Here, the user asks a question and the system retrieves the best-matching answer from curated content. The key distinction: Q&A is primarily retrieval (find an existing answer), while generative chat is primarily generation (compose new text). In enterprise scenarios, Q&A is popular because it is more controllable and easier to validate than free-form generation.
Exam Tip: If the requirement says “respond with answers from our approved documentation” or “use an FAQ,” think Q&A retrieval patterns. If it says “draft a new response” or “converse naturally across many topics,” that leans toward GenAI chat (often via Azure OpenAI) or a hybrid approach.
Common trap: assuming a conversational interface always implies GenAI. The exam can present a chatbot scenario where the best answer is still classic intent/entity extraction and a deterministic workflow (e.g., route a ticket, look up an order status). Another trap: ignoring escalation and fallback. When intent confidence is low or no answer is found, the correct design is often to hand off to a human or request clarification—this is both a user experience and responsible AI expectation.
To select correct answers, watch for keywords: “intents/entities” suggests NLU-style conversational AI fundamentals; “knowledge base/FAQ” suggests Q&A retrieval; “grounded responses” suggests retrieval plus generation (RAG), which you’ll revisit in Section 5.6.
Generative AI workloads are defined by producing new content: text, summaries, explanations, translations, and sometimes code or structured outputs. On the exam, the scenario cues are verbs like draft, rewrite, brainstorm, compose, summarize with a specific tone, or generate variants. These are ideal fits for large language models (LLMs) accessed through Azure OpenAI or productized through Copilot experiences.
Benefits the exam expects you to articulate: improved productivity (first drafts, faster analysis), natural language interfaces (users ask questions instead of navigating complex apps), and adaptability (the same model can handle multiple language tasks via prompting). Generative AI also supports “transformations,” like converting bullet points into a formal email or turning a meeting transcript into action items.
Limitations are heavily tested, especially under the “responsible AI” lens. LLMs can hallucinate (produce plausible but incorrect statements), amplify bias present in training data, and leak sensitive information if prompts or outputs are mishandled. They also require careful evaluation: you don’t “prove correctness” the way you might for deterministic extraction; you validate quality, safety, and grounding.
Exam Tip: When a scenario requires high factual accuracy tied to enterprise data (policies, inventory, customer records), assume you need grounding (connecting the model to trusted sources) rather than asking the model to “just know” the answer. This points toward retrieval-augmented generation concepts discussed later.
Common trap: treating GenAI as a replacement for all NLP. If the question needs consistent labels, compliance scoring, or entity extraction at scale, classic NLP services are often the best match. GenAI excels when the output is flexible language, but it increases governance and validation needs.
AI-900 uses Azure OpenAI terminology in a conceptual way: model families, prompting, and key generation controls. You are not expected to code, but you are expected to understand what terms mean and why they matter in solution selection and responsible deployment.
Tokens are chunks of text used by the model for input and output. Token limits affect how much context you can send (prompt + conversation history + retrieved documents) and how long the response can be. Embeddings are numeric vector representations of text used to measure similarity; they power semantic search and retrieval patterns that “ground” answers. The exam often frames embeddings as the foundation for “find related documents” and “search by meaning, not keywords.”
Prompts are the instructions and context you provide. Good prompts specify role, task, constraints, and desired format. On the exam, prompt engineering is usually assessed through outcomes: does the prompt reduce ambiguity, require citations, or constrain the model to provided content?
Generation controls: temperature and top-p both influence randomness/creativity. Higher temperature (or broader top-p) generally yields more varied outputs; lower values tend to be more deterministic and repeatable. In regulated or customer-facing scenarios (policy answers, support guidance), the safer default is lower creativity to reduce unexpected phrasing and errors.
Exam Tip: If you see “consistent answers” or “reduce variability,” choose lower temperature/top-p. If you see “brainstorm ideas” or “creative copy,” choose higher. The trap is thinking higher creativity improves factuality—it usually does the opposite.
Finally, remember “grounding” vocabulary: the model can be instructed to rely on provided context, but without retrieval it may still hallucinate. This is why Azure OpenAI solutions commonly pair with search/retrieval to anchor responses to enterprise content.
Copilot is the exam’s way of describing productized generative AI assistants embedded into tools users already work in. Rather than building an app from scratch, organizations adopt Copilot experiences to accelerate common tasks: summarizing meetings, drafting emails, generating presentations, answering questions over business content, and automating workflows. On AI-900, you’re tested on recognizing when “use a Copilot” is the most appropriate answer versus “build a custom Azure OpenAI solution.”
Productivity patterns to remember: (1) Draft and refine (first draft → edit with constraints), (2) Summarize and extract actions (notes → decisions, tasks), (3) Ask-and-answer over organizational knowledge (policies, projects), and (4) Automate routine work (connect to business processes through low-code tooling). Copilot is often described as a natural-language front end that orchestrates actions across apps while respecting identity and permissions.
Governance basics appear as exam-friendly terms: data protection, access control, auditability, and responsible AI safeguards. You should assume enterprises require controls around what data is used, who can access it, and how outputs are monitored. Even without deep implementation detail, the exam expects you to prefer solutions that align with organizational compliance, privacy, and safety requirements.
Exam Tip: When the scenario is “employees want AI inside Microsoft productivity tools” (email, documents, meetings) and emphasizes quick adoption, Copilot is often the intended answer. When the scenario is a unique customer-facing app with custom workflows and domain data integration, Azure OpenAI is often the intended answer.
Common trap: treating Copilot as “just a chatbot.” Copilot is typically positioned as an assistant integrated into workflows with governance and enterprise identity; the exam will reward you for linking Copilot to productivity and organizational controls rather than generic chat.
This section ties together what AI-900 tests most: can you describe how to get useful, safer outcomes from GenAI without overstating its capabilities. Three recurring concepts appear in exam language: prompt engineering, retrieval-augmented generation (RAG), and responsible deployment.
Prompt engineering basics: specify the task, the audience, the constraints, and the output format. Strong prompts reduce ambiguity (which reduces hallucinations) and make evaluation easier (because outputs are consistent). Look for “must use bullet points,” “must cite provided policy,” or “if you don’t know, say you don’t know” as exam indicators of good prompting. A common trap is prompts that ask for “the best answer” without giving context—these invite the model to fill gaps with guesses.
RAG/grounding: grounding means anchoring responses to trusted content. RAG is the common pattern: (1) create embeddings for documents, (2) retrieve the most relevant passages by similarity, (3) provide those passages to the model as context, and (4) generate an answer that references that context. On the exam, the right choice often appears when the scenario says “use our internal documentation,” “avoid hallucinations,” “keep answers up to date,” or “don’t retrain the model.” RAG is preferred when data changes frequently because you update the source documents and index rather than retrain the LLM.
Safe deployment: the exam expects you to acknowledge content safety, privacy, and monitoring. You should be ready to describe safeguards such as filtering harmful content, limiting sensitive data in prompts, applying least privilege access to data sources, and adding human review for high-impact use cases. Also remember that evaluation is ongoing—prompt changes, new documents, or new user behaviors can change risk.
Exam Tip: If a question contrasts “fine-tune the model” vs “use grounding/retrieval,” the exam commonly favors grounding/RAG for enterprise knowledge scenarios because it’s faster, cheaper, and keeps content current. Fine-tuning is more appropriate for style/format consistency or specialized behavior—not as a primary method for injecting private facts.
Final mini-review: Classic NLP (sentiment, entities) produces structured insights; conversational patterns (intents, Q&A) focus on routing or retrieving known answers; GenAI (Azure OpenAI) generates and transforms language; Copilot applies GenAI in productivity contexts with governance. Many real solutions combine them—your exam task is to identify which component is doing the “main job” in the scenario and select the service accordingly.
1. A support team receives thousands of customer emails per day. They need to automatically detect the customer’s sentiment and extract key phrases from each email. The output must be consistent and explainable (not free-form generation). Which Azure service is the best fit?
2. A company wants an internal chat experience that answers employee questions using the company’s HR policy documents. They want to reduce hallucinations by ensuring responses are grounded in retrieved content from their documents. Which concept best describes this approach?
3. You are designing a solution to find semantically similar documents (for example, policies related to “parental leave” even when the phrase isn’t used). Which generative AI concept is most directly used to enable semantic similarity search?
4. A business user asks for generative AI capabilities inside Outlook and Teams to summarize long email threads and meetings, and to draft responses using organizational context and governance controls. They want a product experience rather than building custom apps. Which option best fits?
5. A company is deploying an Azure OpenAI chat solution for employees. They must reduce the risk of harmful or inappropriate outputs and enforce safety policies. Which capability best addresses this requirement?
This chapter is your consolidation and execution plan. AI-900 rewards clear thinking more than memorization: identify the workload, map it to the correct Azure AI capability, and avoid distractors that sound “AI-ish” but don’t solve the scenario. You’ll complete two mixed-domain mock blocks, then run a structured answer review that trains you to eliminate distractors quickly. Finally, you’ll build a weak-spot remediation plan aligned to official AI-900 outcomes—AI workloads, machine learning fundamentals, computer vision, NLP, and generative AI (Azure OpenAI + Copilot + responsible AI)—and lock in an exam-day checklist that reduces avoidable mistakes.
As you go, keep a single goal: convert knowledge into repeatable decisions under time pressure. The mock exam is not just a score; it’s a diagnostic tool. Your final review is not re-reading; it’s targeted remediation and confidence building through deliberate practice.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Final rapid review and confidence pass plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Final rapid review and confidence pass plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Treat the mock exam like the real one: quiet environment, single sitting, no notes, and a hard time cap. Your objective is to practice decision-making with enough speed to leave room for review. AI-900 questions are typically short, but the traps are subtle: similar service names, overlapping capabilities (for example, “search” vs “language understanding”), and scenario clues that imply constraints like real-time inference, data residency, or responsible AI requirements.
Use a three-pass pacing method. Pass 1 is “confident picks”: answer everything you can in under a minute and flag anything that requires rereading. Pass 2 is “scenario decoding”: return to flagged items and slow down enough to extract the workload type (vision/NLP/generative/ML) and the key verbs (classify, extract, generate, chat, detect objects). Pass 3 is “risk management”: only change answers when you can name the exact clue that proves the new choice is correct.
Exam Tip: In AI-900, the fastest way to get unstuck is to restate the problem as a workload label. If you can say “this is OCR,” “this is sentiment analysis,” or “this is text generation with safety requirements,” the correct service family usually becomes obvious.
Common pacing trap: overinvesting early in one ambiguous question. Flag it and move on. Many later questions jog memory and reduce uncertainty when you return.
Part 1 is designed to mix AI fundamentals with practical scenario identification. Expect rapid context switches: a question may start with “predict” (machine learning) and then shift to “extract key phrases” (NLP), or mention “images from a factory line” (computer vision) alongside “generate a report” (generative AI). Your job is to separate what is being asked from what is merely background.
AI-900 frequently tests the difference between training and inference. In training, you are building a model (choosing algorithms, evaluating metrics). In inference, you are using a trained model to score new data. If a scenario describes deploying an endpoint, scoring new transactions, or running predictions in production, it’s inference. If it describes splitting data, evaluating accuracy, or reducing overfitting, it’s training/evaluation.
Exam Tip: When you see “evaluate,” “validate,” “test set,” “precision/recall,” or “confusion matrix,” you are in model evaluation territory—not just “using AI.” Questions often reward knowing which metric fits classification vs regression.
Also watch for foundational workload identification: classification vs regression vs clustering. Classification predicts a category (fraud/not fraud). Regression predicts a number (demand forecast). Clustering groups unlabeled items (customer segmentation). A common distractor is presenting regression-like language (“predict”) but the output is clearly categorical.
In mixed scenarios, prioritize the primary objective. If the goal is to extract text from receipts, OCR is the core workload; any downstream summarization is secondary. If the goal is to create marketing copy from a product brief, that is generative AI; sentiment analysis may be an add-on but not the main service selection.
Part 2 leans harder into Azure service selection, including generative AI. The exam expects you to recognize which Azure AI service matches a workload without overengineering. For vision tasks, distinguish between image analysis (labels/tags), object detection, face-related capabilities, and OCR/document extraction. For NLP, separate text analytics (sentiment, entities, language detection) from conversational AI patterns. For generative AI, focus on Azure OpenAI use cases (text generation, summarization, chat, embeddings) and how Copilot fits as an application layer experience.
Service selection traps often use near-synonyms. “Translate text” indicates translation capabilities (language services), not generative AI. “Extract key phrases” is text analytics, not Q&A chat. “Search a knowledge base” may hint at Azure AI Search plus embeddings; however, if the question is only about understanding intent and entities from user utterances, it’s closer to conversational language understanding. Use the action verb and input/output types to decide.
Exam Tip: For generative AI questions, look for words like “generate,” “draft,” “summarize,” “rewrite,” “chat,” or “create embeddings.” If the requirement includes grounding on enterprise data, think retrieval augmentation patterns (often Azure AI Search + Azure OpenAI embeddings) rather than “fine-tune” as a default.
Copilot concepts are tested as “what it is” and “what it is not.” Copilot is typically an integrated assistant experience built on LLMs and orchestration (prompts, plugins/connectors, grounding, and safety). The trap is assuming Copilot equals “a model you train.” In most exam framing, you configure and govern; you do not train the underlying foundation model. Likewise, responsible AI shows up as requirements: content filtering, privacy, human oversight, transparency, and avoiding harmful outputs. If a scenario emphasizes safety constraints, choose options that mention governance controls rather than “increase creativity” or “remove filters.”
Finally, keep an eye on “best choice” questions that implicitly prefer managed services over custom builds. AI-900 is fundamentals: select the simplest managed Azure AI capability that meets the need.
Your score improves fastest when you review answers like an examiner. Don’t just note the correct option—write a one-sentence justification and a one-sentence rejection for each distractor. This builds “discrimination skill,” which is what the test actually measures: can you tell similar concepts apart under time pressure?
Use a consistent four-step review method. Step 1: identify the workload (ML, vision, NLP, generative). Step 2: identify the task subtype (classification vs regression; OCR vs object detection; sentiment vs entity extraction; generation vs embeddings). Step 3: identify constraints (real-time inference, structured vs unstructured input, safety/governance). Step 4: verify the service mapping aligns with those constraints.
Exam Tip: If you can’t explain why three options are wrong, you don’t truly “own” the concept yet. In the real exam, uncertainty usually comes from not being able to eliminate distractors confidently.
Common distractor patterns include:
When you miss a question, label the miss type (misread, concept gap, service confusion, or overthinking). Your remediation plan in the next section should target the miss type, not just the topic.
After both mock parts, categorize every miss and every “guessed but correct” item. Guesses are hidden weaknesses: they tend to fail under exam stress. Map them to the course outcomes (which align to exam domains) and create a short remediation loop for each.
Domain 1: AI workloads and solution scenarios. Remediate by drilling workload keywords and outputs. Practice converting a scenario into: input type → task verb → output type. Trap: “AI” vs “automation.” Not every automation needs AI; the exam may bait you with simple rule-based cases.
Domain 2: ML fundamentals on Azure (training vs inference, evaluation). Remediate by reviewing metric-to-task mapping: accuracy/precision/recall for classification; MAE/MSE/RMSE for regression; overfitting signals and train/validation/test roles. Trap: thinking higher accuracy always wins—imbalanced datasets often require precision/recall focus.
Domain 3: Computer vision workloads. Remediate by separating OCR/document extraction from image classification and object detection. Trap: “detect text” vs “analyze image.” If the output is extracted strings and fields, it’s OCR/document intelligence style thinking.
Domain 4: NLP workloads (text analytics and conversational concepts). Remediate by practicing the boundaries: sentiment/key phrases/entities/language detection vs intent-based conversations. Trap: equating “chat” with “NLP service selection”—sometimes the scenario only needs analytics, not a bot.
Domain 5: Generative AI on Azure (Azure OpenAI, responsible AI, Copilot). Remediate by outlining: generation vs embeddings vs retrieval augmentation; and responsible AI controls (content filtering, monitoring, human oversight, privacy). Trap: defaulting to fine-tuning when grounding via retrieval is safer and faster for most enterprise Q&A use cases.
Keep remediation tight: 20–30 minutes per weak domain, then re-attempt similar items or review notes until you can reject distractors confidently.
On exam day, your goal is to protect focus and reduce preventable errors. Start with environment control: reliable network, quiet space, and all required identification if proctored. Close unnecessary apps, silence notifications, and ensure your display settings are comfortable for reading—fatigue leads to misreading “best” vs “least” questions.
Use the same three-pass strategy you practiced. Pass 1: answer fast and flag. Pass 2: resolve flags by mapping to workload and service boundaries. Pass 3: audit for misreads, especially negatives (“NOT”), superlatives (“most appropriate”), and scope (“which service,” “which feature,” “which principle”).
Exam Tip: Do not change an answer in the final minutes unless you can point to a specific requirement word in the prompt that contradicts your original choice. “A feeling” is not evidence.
Last-minute review should be lightweight: revisit your personal “trap list” (common confusions you identified during weak-spot analysis) and the high-frequency mappings: classification/regression/clustering; training vs inference; OCR vs image analysis; sentiment/entities vs conversational intent; Azure OpenAI generation/embeddings and Copilot as an experience layer with responsible AI guardrails. Walk in with a plan, execute your passes, and let disciplined elimination do the work.
1. You are reviewing results from a full AI-900 mock exam. A user story reads: “We need to extract key phrases and detect the language from customer support tickets.” Which Azure AI capability is the best fit for this workload?
2. A company wants a chatbot that can draft answers using the company’s policy documents. They want to reduce hallucinations by grounding responses in their own content. Which approach best matches Azure OpenAI recommended patterns for this scenario?
3. During weak-spot analysis, you notice you often confuse classification and regression questions. Which scenario describes a classification workload?
4. You’re applying an exam-day checklist strategy. A question asks: “Select the Azure service that detects objects in images and returns bounding boxes.” Which service should you select?
5. Your final rapid review focuses on responsible AI. A team is deploying a generative AI assistant using Azure OpenAI and wants to reduce the risk of generating harmful content. What is the most appropriate built-in capability to apply?