AI Certification Exam Prep — Beginner
Learn AI-900 essentials fast with clear Azure examples and exam-style practice.
“AI-900 in 7 Days: Azure AI Concepts Explained Simply” is a beginner-friendly exam-prep course for the Microsoft Azure AI Fundamentals (AI-900) certification. If you have basic IT literacy but feel overwhelmed by AI terminology, this course breaks the exam down into clear daily targets, plain-language explanations, and realistic practice questions that mirror the style of the exam.
This course is organized to match the official AI-900 domains and the way Microsoft expects you to think on exam day. You’ll learn how to identify the right Azure capability for a scenario, understand key AI and machine learning concepts, and avoid common distractors.
Chapter 1 sets you up for success with exam orientation: registration steps, scoring expectations, question formats, and a practical 7-day study strategy. Chapters 2 through 5 each focus on one or two exam domains, moving from core concepts to “choose the best service/approach” decision-making—the skill that shows up repeatedly on AI-900. Chapter 6 finishes with a full mock exam experience, weak-spot analysis, and an exam-day checklist so you know exactly what to do in the final hours before your test.
You can begin immediately and follow the plan day-by-day. If you’re new to certification prep, the course shows you how to study efficiently (active recall, spaced repetition, and targeted review) so you spend time where it actually increases your score. When you’re ready to begin, use Register free to access the platform. If you want to compare learning paths, you can also browse all courses.
By the end, you will be able to explain each AI-900 domain in simple terms, select the right Azure AI option for common scenarios, and complete a full mock exam with a clear review process—so you walk into the Microsoft AI-900 exam calm, prepared, and confident.
Microsoft Certified Trainer (MCT)
Jordan Whitaker is a Microsoft Certified Trainer who helps beginners pass Microsoft fundamentals exams through clear explanations and hands-on Azure examples. He has designed certification prep programs for Azure, data, and AI tracks with a focus on exam objectives and common pitfalls.
AI-900 (Azure AI Fundamentals) is designed to confirm that you can recognize common AI workloads, describe how machine learning works at a high level, and map real-world scenarios to the right Azure AI services. This chapter orients you to the exam format, how Microsoft defines the skills measured, and how to build a tight 7-day plan that emphasizes recall practice and review loops. Treat AI-900 less like a memorization test and more like a “service selection and concept clarity” exam: most items reward choosing the best fit given a scenario, constraints, and wording.
Over the next week, your job is to (1) learn the vocabulary that appears on the test, (2) practice translating scenarios into solution types (ML vs vision vs language vs generative AI), and (3) get comfortable with how Microsoft writes distractors—options that sound plausible but don’t match the workload or service boundary. You will also set your timeline, schedule the exam through Pearson VUE, and understand scoring and retakes so exam day feels routine rather than risky.
Practice note for Understand AI-900 format, question types, and timing: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam (Pearson VUE) and set your timeline: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scoring, skills measured, and how Microsoft updates objectives: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a 7-day plan: daily targets, recall practice, and review loops: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand AI-900 format, question types, and timing: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam (Pearson VUE) and set your timeline: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scoring, skills measured, and how Microsoft updates objectives: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a 7-day plan: daily targets, recall practice, and review loops: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand AI-900 format, question types, and timing: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam (Pearson VUE) and set your timeline: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 validates foundational literacy in Azure AI—enough to participate in solution discussions and make correct “what should I use?” decisions. On the exam you are not expected to build models from scratch, tune hyperparameters, or write production code. Instead, you must recognize AI workload categories and match them to appropriate Azure services and approaches. This ties directly to the course outcomes: describing AI workloads (e.g., prediction, classification, anomaly detection), explaining core ML principles (training vs inference; features/labels; supervised vs unsupervised), mapping computer vision scenarios to Azure AI Vision, mapping NLP and conversation scenarios to Azure AI Language and Speech, and describing generative AI fundamentals and Azure OpenAI use cases.
The most common exam behavior is scenario-to-solution mapping. Example patterns: “identify objects in an image” maps to vision; “extract key phrases and detect sentiment from reviews” maps to language; “transcribe audio” maps to speech; “generate marketing copy given a prompt” maps to generative AI and Azure OpenAI. The trap is overthinking implementation details when the exam is testing conceptual fit. If a question asks for “a service” and your answer is an algorithm (like logistic regression), it’s likely wrong unless the question is explicitly about ML concepts rather than Azure products.
Exam Tip: When you see words like classify, predict, forecast, recommend, think machine learning. When you see detect, recognize, analyze images, think vision. When you see sentiment, entities, language understanding, translation, speech-to-text, think NLP/speech. When you see prompt, completion, embeddings, chat, generate, think generative AI.
Microsoft organizes AI-900 around a “skills measured” outline that is periodically updated. Your first job is to study from the current official objectives, not a blog post from last year. The domains typically cover: (1) AI workloads and considerations, (2) fundamental principles of machine learning on Azure, (3) features of computer vision workloads on Azure, (4) features of NLP workloads on Azure, and (5) features of generative AI workloads on Azure. Even when the names shift slightly, the testing intent stays consistent: identify the right solution type and understand what each Azure AI service is for.
In practice, each domain has two layers: concept vocabulary and service mapping. Concept vocabulary includes terms like training, inference, overfitting, evaluation metrics (at a high level), responsible AI considerations, and the difference between supervised and unsupervised learning. Service mapping includes knowing what problems Azure AI services solve (for example, Azure AI Vision for image analysis, Azure AI Language for text analytics, Azure AI Speech for speech-to-text/text-to-speech, and Azure OpenAI for generative use cases). A frequent distractor is mixing up “Vision” and “Language” features when the input type is clear (image vs text). Another is choosing a generalized “Azure Machine Learning” answer when the question is clearly about a prebuilt cognitive service capability.
Exam Tip: If the scenario is a common, prebuilt capability (OCR, sentiment analysis, key phrase extraction, face detection, transcription), assume the exam wants the relevant Azure AI service, not a custom model approach. Reserve Azure Machine Learning for scenarios that require custom training, model management, or more control than a prebuilt API provides.
Because Microsoft updates objectives, build a habit now: check the “Skills measured” page before you start Day 1 and again a day or two before your exam. Objective changes are usually additive (new GenAI emphasis, renames), and a last-minute read can prevent surprise terminology on exam day.
You’ll register through Microsoft’s certification portal and schedule via Pearson VUE. The operational goal is simple: pick an exam date that matches your 7-day plan and eliminates decision fatigue. AI-900 is approachable, but deadlines matter—without a scheduled date, “studying” often becomes vague reading rather than deliberate practice. Choose either an online proctored exam or a test center. Online is convenient but stricter on room setup; test centers reduce environment issues.
For identity and policy compliance, plan ahead. You typically need government-issued photo ID that matches your registration name. Mismatched names are a preventable failure mode—fix your profile before scheduling. For online proctoring, expect rules around a clear desk, no extra monitors, and no audible reading. Your system will run a check for webcam, mic, and network stability; do this early to avoid last-minute technical disruptions. Pearson VUE policies also cover breaks and what you may have on your desk; assume “nothing” unless explicitly allowed.
Exam Tip: Schedule the exam for Day 8 or late Day 7 evening. This gives you 7 full days of preparation and preserves a buffer for rescheduling if life intervenes. Avoid scheduling immediately after a workday if you know your focus will be low—AI-900 rewards careful reading of scenarios and option wording.
Also note that Microsoft can update exam objectives, but your exam appointment stays valid. Your responsibility is to align your preparation to the current objectives during your study window. The earlier you schedule, the more disciplined your daily targets become.
Microsoft certification exams typically use a scaled scoring model. You don’t need to compute percentages; you need consistent accuracy across domains. The exam may include different question types (for example, multiple choice, multiple response, drag-and-drop/matching, case-style scenario sets). Time pressure is real mainly because of careful reading, not because of lengthy math. Your pacing goal is to keep moving and avoid getting stuck proving yourself “right” on one item.
Passing expectations: treat 700 (on a 1000 scale) as the target threshold and build a buffer above it by strengthening weak domains. Many candidates fail not due to lack of intelligence, but due to (1) confusing service boundaries (Vision vs Language vs Speech), (2) mixing ML concepts (supervised vs unsupervised), and (3) misreading “best answer” prompts that prioritize operational simplicity (prebuilt service) over custom build (Azure Machine Learning).
Exam Tip: Don’t aim for perfection on obscure edge cases. Aim for dependable performance on the high-frequency mapping items: vision vs language vs speech vs genAI, and the difference between training a model and using a prebuilt capability.
Retake strategy should be part of your plan even if you expect to pass. If you do not pass, immediately write down which domains felt unfamiliar and which distractors fooled you. Then re-study using the official objectives and do targeted practice. The trap is repeating the same study method (passive reading) and expecting a different result. Your retake plan should emphasize recall practice and scenario mapping, not more notes.
Your 7-day plan should be structured around learning, recall, and review loops. Beginners often read content once and feel confident—until they face scenario-based questions. Instead, use spaced repetition: revisit key concepts repeatedly with increasing intervals. Your notes should be short, decision-focused, and phrased as “If you see X, choose Y.” Build a one-page “service map” that lists workload → Azure service → typical tasks. This is the fastest path to exam readiness because AI-900 emphasizes identification and selection.
Recommended 7-day structure: Days 1–2 focus on AI workloads and ML fundamentals (training vs inference; supervised vs unsupervised; regression vs classification vs clustering; responsible AI). Days 3–4 focus on Vision and Language/Speech, with a strong emphasis on inputs/outputs: image/video vs text vs audio, and what each service returns. Day 5 focuses on generative AI fundamentals (prompts, tokens, embeddings at a conceptual level, typical use cases) and Azure OpenAI positioning. Day 6 is mixed review with targeted drills on weak areas. Day 7 is full review: service map recitation, revisit traps, and light practice to stay fresh without burning out.
Exam Tip: Use “active recall” daily: close your notes and verbally explain the difference between classification and regression, or between OCR and image captioning, or between speech-to-text and text-to-speech. If you can’t explain it without looking, you don’t own it yet.
Labs and demos help even at fundamentals level. If possible, use quick Azure portal or documentation walkthroughs to see what each service does and how it’s described. The exam often uses Microsoft’s phrasing; reading official docs headings and feature lists can improve recognition. Keep lab time time-boxed: the goal is to cement terminology and capabilities, not to build a full project.
AI-900 questions are designed to test whether you can identify the right workload and the simplest correct Azure option. Your first step on every question is classification: What is the input (text, image, audio, tabular data)? What is the desired output (label, score, extracted entities, transcript, generated text)? Then choose the service or concept that directly matches. Many wrong answers are “near neighbors” (related but not correct), so practice reading for the exact task.
Common distractor patterns: (1) choosing Azure Machine Learning when a prebuilt Azure AI service is sufficient; (2) choosing a vision feature for a language problem or vice versa; (3) confusing generative AI (creating new content) with classic NLP analytics (extracting information from existing text). Another trap is ignoring the word “best.” If multiple options could work, “best” often means fastest to implement, least custom training, or most managed—typically a prebuilt service.
Exam Tip: Underline (mentally) constraint words: “real-time,” “custom,” “no training data,” “extract,” “generate,” “summarize,” “detect language,” “transcribe.” These words usually eliminate half the options.
Timing strategy: do a first pass answering what you know confidently. If unsure, eliminate options based on input/output mismatch. If two remain, pick the one that aligns with the workload category and is most directly described by Microsoft product language. Avoid changing answers without a clear reason; second-guessing often swaps a correct service match for a plausible-sounding distractor. Your daily practice should include reviewing why wrong options are wrong—because that is exactly how you immunize yourself against distractors on exam day.
1. You are starting a 7-day study sprint for AI-900. Your goal is to maximize your score by focusing on what the exam rewards most. Which approach best aligns with AI-900’s typical question style and scoring? A. Memorize as many Azure AI service names and SKUs as possible. B. Practice mapping scenarios to the correct AI workload and Azure AI service boundary. C. Focus primarily on implementing end-to-end machine learning pipelines with code samples.
2. A student is scheduling AI-900 and wants to reduce risk on exam day. Which action best reflects the recommended process described in the chapter? A. Schedule the exam through Pearson VUE and set a fixed timeline for your 7-day plan. B. Schedule the exam only after completing all practice tests to avoid rescheduling fees. C. Register through the Azure portal because AI-900 is an Azure service exam.
3. Your team maintains an internal AI-900 prep guide. Microsoft periodically changes what appears on the exam. What is the most reliable source to verify whether the objectives have been updated? A. The course’s chapter notes because they are fixed for the current cohort. B. The official Microsoft “skills measured” / exam skills outline for AI-900. C. Community forum posts summarizing what recent test-takers saw.
4. You have only 7 days to prepare. Which daily routine best matches a high-retention plan for AI-900 described in the chapter? A. Read all material once, then take a single full practice exam on day 7. B. Set daily targets and include recall practice with review loops (spaced review of missed concepts). C. Watch videos at 2× speed and skip practice questions to save time.
5. A company wants to brief new hires on what to expect from the AI-900 exam experience. Which statement best reflects the chapter’s guidance about question types and timing? A. Expect only long case studies; time management is not a concern because all questions are open-book. B. Expect a mix of question styles (including scenario-based items) and plan pacing because time is limited. C. Expect hands-on lab tasks in Azure that require deploying resources during the exam.
This chapter targets a core AI-900 skill: recognizing what kind AI problem you’re looking at and selecting the right approach or Azure capability. The exam is not asking you to design neural networks from scratch; it tests whether you can map scenarios to workload categories (vision, language, predictions, recommendations), understand the high-level differences between machine learning (ML), deep learning (DL), and generative AI (GenAI), and apply Responsible AI basics to outcomes.
As you read, practice translating everyday business statements (“we need to flag unusual transactions,” “summarize customer calls,” “identify objects in a photo”) into workload keywords that Azure and the exam use (classification, anomaly detection, computer vision, speech-to-text, retrieval, and generation). That translation skill is what turns long scenario questions into quick, confident answers.
Exam Tip: In AI-900, most “hard” questions become easy when you name the workload precisely. Once you say “this is classification” or “this is OCR,” the service choice is usually straightforward.
Practice note for Recognize AI workload categories and real-world scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate ML, deep learning, and generative AI at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI basics to workloads and outcomes: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: AI workload identification and concept checks: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize AI workload categories and real-world scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate ML, deep learning, and generative AI at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI basics to workloads and outcomes: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: AI workload identification and concept checks: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize AI workload categories and real-world scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate ML, deep learning, and generative AI at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI on the AI-900 exam is framed as systems that perform tasks typically associated with human intelligence: perceiving (vision/speech), understanding language, learning patterns from data, and generating content. The key practical distinction is between traditional software logic (explicit rules) and AI workloads (models learn from examples or large-scale pretraining).
Traditional logic works best when rules are stable and fully known: “If customer age < 18, block purchase.” AI workloads are better when rules are hard to enumerate or change frequently: “Detect fraudulent behavior,” “Recognize items in images,” or “Predict demand next week.” In AI terms, your inputs (features like transaction amount, time, device) map to outputs (fraud/not fraud, forecast value, detected objects) through a model that generalizes from training data.
AI-900 often checks whether you understand that AI is probabilistic. A model gives a best guess, often with confidence. Traditional code is deterministic. That matters for real-world deployment: you monitor model drift, handle uncertainty, and evaluate performance.
Exam Tip: If a scenario says “define rules” or “fixed logic,” it’s usually not an AI workload. If it says “learn patterns,” “make predictions,” “recognize,” “classify,” or “understand,” it is.
Common trap: assuming “automation” equals AI. Automation can be simple scripting or workflow tools. The exam typically signals AI by referencing training data, labeled examples, or unstructured inputs (text, images, audio) where rules are messy.
This section directly supports “Recognize AI workload categories and real-world scenarios.” AI-900 expects you to identify common ML workload types by the shape of the output.
Prediction usually means a numeric value (regression) or a future estimate: forecasting sales, predicting temperature, estimating time-to-failure. Clues include “how many,” “how much,” “forecast,” “estimate,” and “next month.”
Classification assigns a category label. Binary classification is yes/no (approve/deny, fraud/not fraud). Multi-class is among several labels (route ticket to billing/technical/sales). Clues include “which category,” “label,” “classify,” “is this spam.”
Detection is overloaded on exams. It can mean: (1) anomaly detection (unusual transactions, sensor spikes), (2) object detection in images (find objects and bounding boxes), or (3) entity detection in text (names, organizations, PII). Read the context: numeric time series suggests anomalies; images suggest object detection; text suggests entity extraction.
Recommendation predicts preferences: “customers who bought X also bought Y,” “rank products,” “personalized content.” The output is often an ordered list. The scenario usually mentions user-item history or personalization.
Exam Tip: When two answers look plausible, pick the one that matches the output type: number (prediction), label (classification), location/flag (detection), ranked list (recommendation).
Common traps: confusing classification vs detection (classification labels the whole item; object detection locates items within an image) and confusing prediction vs recommendation (recommendation is a ranking/personalization problem, not just “predict a value”).
AI-900 tests whether you can choose between prebuilt Azure AI services and custom machine learning on Azure. Think in families: Azure AI services provide ready-made models via APIs; Azure Machine Learning supports training, tuning, and deploying your own models.
Azure AI services (often referred to as “Cognitive Services” historically) cover common workloads with minimal ML expertise. Examples you should recognize: computer vision capabilities (image analysis, OCR, detection), language capabilities (sentiment, key phrases, entity recognition), speech capabilities (speech-to-text, text-to-speech, translation), and content safety. These are ideal when the task is common and you can accept general-purpose behavior.
Custom ML with Azure Machine Learning fits when you need a model tailored to your data: predicting churn from your customer behavior, a specialized risk model, or domain-specific classification. Azure Machine Learning provides experiment tracking, training pipelines, model registry, and deployment endpoints.
High-level deep learning fits within both: many prebuilt services use deep learning internally, and you can also train deep learning models yourself (especially for vision or language) if prebuilt accuracy is insufficient or you need domain control.
Exam Tip: If the scenario says “train a model,” “use your own labeled data,” “custom algorithm,” or “full control,” lean toward Azure Machine Learning. If it says “extract text,” “detect faces/objects,” “translate,” “analyze sentiment,” lean toward Azure AI services.
Common trap: choosing Azure Machine Learning when the scenario is clearly an off-the-shelf capability (for example, OCR from images). The exam rewards using the simplest appropriate service rather than building everything from scratch.
“Apply responsible AI basics to workloads and outcomes” is a recurring exam objective. AI-900 typically checks recognition-level understanding: you should identify which Responsible AI principle is at risk in a scenario and what type of mitigation is implied.
Fairness is about avoiding discriminatory outcomes across groups (for example, different approval rates not explained by legitimate factors). Watch for scenarios involving hiring, lending, insurance, or criminal justice—these often point to fairness evaluation and bias mitigation.
Reliability and safety focuses on consistent performance under expected conditions and resilience to edge cases. Clues include “works in the lab but fails in production,” “rare conditions,” “model drift,” or “safety-critical decisions.”
Privacy concerns appropriate collection, use, retention, and sharing of data. Scenarios with personal data, voice recordings, or medical info suggest privacy requirements such as minimization and access controls.
Security includes protecting models and data from attacks (prompt injection in GenAI contexts, data poisoning, endpoint abuse) and ensuring proper authentication/authorization. If the scenario mentions “malicious inputs,” “data tampering,” or “unauthorized access,” it’s security.
Transparency is about explainability and clarity that users are interacting with AI, plus understandable reasons for outputs when needed. If a user asks “why was I denied?” or regulators require reasoning, transparency is central.
Exam Tip: Don’t over-engineer answers. AI-900 typically wants the principle name matched to the scenario. If the prompt highlights sensitive demographics, it’s usually fairness; if it highlights personal data handling, it’s privacy; if it highlights attacks, it’s security.
Common trap: mixing privacy and security. Privacy is about appropriate use of personal data; security is about protection from threats. You often need both in real solutions, but exam questions typically emphasize one.
This decision point appears in many scenario questions: “Which should you use?” The exam expects pragmatic reasoning: time-to-value, data availability, customization needs, and operational complexity.
Use prebuilt AI services when the capability is common and broadly applicable (OCR, speech-to-text, translation, basic image tagging, sentiment analysis). Benefits: fast implementation, no model training, managed scaling, and typically strong baseline performance.
Train your own model (via Azure Machine Learning) when: (1) you have unique data and a specific target (e.g., predicting failures for your equipment), (2) you need custom labels or domain-specific categories, (3) prebuilt services don’t meet accuracy or compliance needs, or (4) you require control over features, evaluation metrics, and retraining cadence.
Generative AI adds a hybrid option: you may use a foundation model (Azure OpenAI) and then ground it with enterprise data (retrieval) or customize behavior with prompts and safety filters—without “training” a full model. AI-900 stays high level, but it does test that GenAI is often consumed as an API service rather than built from scratch.
Exam Tip: If the scenario mentions “lack of labeled data,” prefer prebuilt services or GenAI with retrieval. If it mentions “large labeled dataset” and “custom outcome,” prefer Azure Machine Learning.
Common trap: assuming deep learning is always required. Many business predictions are classic ML; the exam rewards choosing the right workload, not the fanciest technique.
This section supports your “Practice set: AI workload identification and concept checks” by teaching how to think, not by drilling questions. On AI-900, you win by mapping nouns and verbs in the scenario to the workload and then to the service family.
Start with the input type: images/video usually implies computer vision (image analysis, OCR, object detection). Text implies NLP (language detection, key phrases, sentiment, entities). Audio implies speech (speech-to-text, translation, speaker scenarios). Tabular/time-series often implies ML predictions, classification, or anomaly detection. If the task is “write,” “summarize,” “chat,” or “generate,” think generative AI (Azure OpenAI).
Then identify the output: label, number, ranked list, extracted fields, bounding boxes, or generated text. This is where many terminology pitfalls live: “detection” could be anomalies, objects, or entities; “classification” could be document categorization or image labeling; “prediction” could be regression or forecasting.
Exam Tip: Watch for distractor answers that name an algorithm (like “neural network”) when the question is really about a service category (Azure AI services vs Azure Machine Learning) or a workload type (classification vs regression). AI-900 is concept-and-service oriented.
Another common pitfall is confusing GenAI with traditional NLP. Sentiment analysis and key phrase extraction are analysis tasks (Azure AI Language). Writing an email draft, summarizing a meeting, or answering questions in natural language is generation (Azure OpenAI). If the scenario emphasizes “create new content,” that’s GenAI; if it emphasizes “extract insights from existing text,” that’s NLP analysis.
Finally, apply Responsible AI as a cross-check. If the scenario affects people (credit, hiring, healthcare), ensure your mental answer includes fairness/transparency considerations—AI-900 may include an option that explicitly references a responsible principle, and that can be the intended correct choice.
1. A retail company wants to automatically route incoming customer emails into one of four queues: Billing, Shipping, Returns, or Technical Support. Which AI workload is this scenario describing?
2. A bank has years of labeled transaction data, including whether each transaction was fraudulent. The bank wants to predict the probability that a new transaction is fraudulent. Which workload best fits this requirement?
3. A call center wants to convert recorded customer calls into text and then produce a short summary of each call for agents. Which pairing of AI capabilities best matches this end-to-end solution?
4. You are explaining model types to a colleague. Which statement correctly differentiates machine learning (ML), deep learning (DL), and generative AI (GenAI) at a high level?
5. A company deploys an AI model to screen job applications. An internal review finds that qualified applicants from a certain demographic are being rejected at a higher rate. Which Responsible AI principle is most directly being addressed when investigating and mitigating this issue?
On AI-900, “machine learning” is tested as a set of clear fundamentals: what data looks like (features and labels), what the model does during training vs inference, which task type fits a scenario (regression/classification/clustering), and how to interpret basic evaluation results without overthinking. The exam is not trying to make you build models from scratch; it’s checking whether you can match a business problem to the right ML approach and the right Azure service choices.
Think of ML as a repeatable workflow: you collect data, represent it as features (input columns), optionally define a label (the target you want to predict), train a model that learns patterns, and then deploy or use it for inference (making predictions on new data). A common trap is mixing up features and labels in scenario questions. Features describe what you know at prediction time; labels are what you’re trying to predict.
This chapter connects those fundamentals to Azure Machine Learning (Azure ML) and to the exam’s expectation: you should be able to pick the simplest correct solution type, identify common pitfalls like overfitting/underfitting, and recognize which metrics make sense for a given task.
Practice note for Learn core ML concepts: features, labels, training, inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose ML task types: regression, classification, clustering: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand model evaluation and overfitting/underfitting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML fundamentals and Azure ML choices: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Learn core ML concepts: features, labels, training, inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose ML task types: regression, classification, clustering: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand model evaluation and overfitting/underfitting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML fundamentals and Azure ML choices: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Learn core ML concepts: features, labels, training, inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose ML task types: regression, classification, clustering: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 frequently frames scenarios around whether you have labeled outcomes. That single detail tells you whether you are in supervised or unsupervised learning. In supervised learning, your training data includes both features (inputs) and labels (the correct outputs). The model learns a mapping from inputs to outputs. In unsupervised learning, you only have features; the goal is to discover structure (groups, patterns, or anomalies) without predefined “correct” answers.
Common supervised examples: predicting house prices (label is price), predicting whether a transaction is fraud (label is fraud/not fraud), predicting customer churn (label is churn/no churn). Common unsupervised examples: segmenting customers into groups by behavior (no label, you’re discovering clusters), exploring topic groupings in documents, or detecting unusual patterns when you don’t have historical “anomaly” labels.
Exam Tip: Look for wording like “historical records include the outcome” or “known results” (supervised) versus “want to group,” “discover segments,” “no labeled data,” or “unknown categories” (unsupervised). If the scenario asks to “predict” a known target, it’s almost always supervised.
Another testable distinction is how the model is used. During training, the model learns from historical data. During inference, it applies what it learned to new data. A common trap is thinking inference requires labels. It does not—labels are typically absent at inference time because the model’s job is to produce the predicted label.
Once you identify supervised vs unsupervised, AI-900 expects you to choose the right task type. The three core types you’ll see most often are regression, classification, and clustering.
Regression is supervised learning where the label is a number (continuous value). Examples: forecast energy usage, predict delivery time in minutes, estimate sales revenue, predict temperature. The most common trap is confusing “number of categories” with regression. If the output is numeric but represents a category code (for example, 1/2/3 meaning low/medium/high), that’s still classification because the numeric value is just an encoding.
Classification is supervised learning where the label is a category. This can be binary (yes/no) or multi-class (A/B/C). Examples: approve/deny a loan, classify email as spam/not spam, predict which product a customer will buy next from a set of options. If the question says “probability that…” it still points to classification; models often output probabilities for each class.
Clustering is unsupervised learning where you group similar items. Examples: customer segmentation, grouping news articles by similarity, organizing products into groups based on purchase patterns. Clustering is not about predicting a pre-labeled class; it’s about discovering groupings that may later be interpreted by humans.
Exam Tip: Decide based on the output: numeric value → regression; category label → classification; “group/segment/find clusters without labels” → clustering. Then confirm by scanning for label availability: if the scenario never mentions known outcomes, it’s unlikely to be supervised.
AI-900 doesn’t require deep MLOps, but it does test that you understand the basic training pipeline and why each stage exists. Most ML failures are pipeline failures: poor data quality, leakage, or evaluating on the wrong dataset.
Start with data preparation: cleaning missing values, encoding categorical fields, normalizing/standardizing numeric fields when needed, and ensuring your features reflect what will be available at inference time. A classic exam trap is data leakage: including a feature that wouldn’t exist at prediction time (for example, “final invoice amount” to predict “whether the customer paid”). Leakage inflates accuracy during training and fails in real usage.
Next comes data splitting. You typically split into training and test sets, and often also use a validation set. The training set fits the model. The validation set helps tune choices (hyperparameters, feature selection, thresholds). The test set is held back to estimate real-world performance after you’ve made decisions.
Then training produces a model. Validation checks performance while you iterate. Testing is your final check. Overfitting and underfitting show up here: overfitting means excellent training performance but weak validation/test performance (model memorizes noise); underfitting means poor performance everywhere (model is too simple or features don’t capture signal).
Exam Tip: When asked how to improve generalization, think “reduce overfitting”: add more data, simplify the model, add regularization, or improve feature quality. When asked about underfitting, think “increase capacity” or “add better features.” Also, if a question mentions “evaluate on the training data,” that’s usually a red flag—the test set is the proper unbiased estimate.
Metrics are a favorite AI-900 topic because they reveal whether you understand what “good performance” means for different tasks. The exam typically stays at the level of interpreting results and choosing an appropriate metric rather than calculating it by hand.
For classification, accuracy is the proportion of correct predictions overall. The trap: accuracy can be misleading with imbalanced classes (for example, 99% non-fraud). In those cases, the exam expects you to consider precision and recall. Precision answers: “When the model predicts positive, how often is it right?” Recall answers: “Of all actual positives, how many did we catch?” Fraud detection often values recall (catch more fraud), while scenarios with costly false positives may value precision.
A confusion matrix summarizes predictions vs actuals using true positives, false positives, true negatives, and false negatives. Even if you’re not asked to compute values, you should recognize how errors differ: false positives trigger unnecessary actions; false negatives miss the thing you care about.
For regression, common metrics include MAE (mean absolute error) and MSE (mean squared error). MAE is more interpretable (average absolute difference), while MSE penalizes large errors more heavily because of squaring. If the scenario emphasizes “avoid big misses,” MSE/RMSE often fits better conceptually.
Exam Tip: Match metric to risk: imbalanced classification → precision/recall (and confusion matrix for error types); numeric prediction → MAE/MSE. If a question tries to push accuracy for a rare-event problem, that’s often the wrong choice.
AI-900 expects a high-level understanding of Azure Machine Learning (Azure ML) as the platform service for building, training, and deploying ML models. The exam focus is “what is it for” and “what are its key components,” not detailed administration.
An Azure ML workspace is the top-level container that organizes your ML assets: data references, experiments, models, endpoints, and associated settings. If the question asks where you manage and track ML artifacts, the workspace is the anchor concept.
Datasets/data assets (terminology can vary across Azure ML studio updates) represent your data sources used in training and evaluation. The key idea: Azure ML helps you register and reuse data references consistently across experiments.
Compute refers to the resources used to run jobs—commonly compute instances for interactive work and compute clusters for scalable training. For the exam, remember: you don’t need to run training on your laptop; Azure ML provides managed compute options that can scale.
Azure ML supports two common build experiences: Designer (a drag-and-drop, no/low-code interface for pipelines) and code-first development (Python with SDK/CLI). Designer is often positioned for rapid prototyping or when you want visual pipeline assembly; code-first is common for custom training, automation, and advanced scenarios.
Exam Tip: If a scenario emphasizes “no code,” “visual,” or “drag-and-drop pipeline,” choose Azure ML Designer. If it emphasizes “custom training script,” “SDK,” “automate,” or “full control,” choose code-based Azure ML. Also, don’t confuse Azure ML with Azure AI services (prebuilt models). Azure ML is for building/training your own models; Azure AI services are for using Microsoft-provided pretrained capabilities.
On AI-900, the fastest way to earn points is to apply a consistent decision process. First, identify whether you have labeled outcomes. If yes, supervised; if no, unsupervised. Second, determine the output type: numeric (regression) vs categorical (classification) vs “grouping” (clustering). Third, choose the metric that matches the task and the business risk.
When interpreting metrics, watch for imbalanced data traps. If the positive class is rare, a model can show high accuracy while being useless. In those scenarios, focus on precision/recall and the confusion matrix to understand false positives vs false negatives. If the scenario mentions that missing a positive case is costly (for example, fraud slipping through, safety incidents, critical disease screening), prioritize recall concepts. If the scenario emphasizes that acting on a false alarm is expensive (for example, manual reviews, blocking legitimate customers), precision becomes more relevant.
For regression, interpret error metrics in plain language. MAE answers “on average, how far off are we?” MSE emphasizes large errors. If you see a statement like “a few very wrong predictions are unacceptable,” that nudges you toward squared-error thinking.
Finally, connect approach to Azure choices: if the prompt describes building and training a custom model with your own data, Azure ML is the right platform. If it emphasizes a visual workflow and minimal coding, Designer is the likely best fit. If it stresses repeatable automation or custom scripts, code-first Azure ML is the safer match.
Exam Tip: Eliminate answers that mismatch the problem type before debating services. For example, don’t pick clustering when the question asks to predict a known label, and don’t pick accuracy as the only metric for rare-event detection. The exam rewards correct alignment more than advanced nuance.
1. A retail company wants to predict the total sales amount for each store next month based on historical monthly sales, local promotions, and store size. Which machine learning task type should you choose?
2. You are designing a model to predict whether a customer will churn. Your dataset contains columns: CustomerId, TenureMonths, MonthlySpend, SupportTicketsLast30Days, and Churned (Yes/No). Which column is the label?
3. An insurance provider trains a model that performs very well on the training data but performs significantly worse on new, unseen policy applications. Which issue is most likely occurring?
4. A company has customer purchase history with no existing categories. They want to group customers into segments based on purchasing behavior to target marketing campaigns. Which machine learning approach best fits this requirement?
5. You train a model in Azure Machine Learning using historical labeled data. Later, the deployed model receives new records and outputs predictions. What is the correct term for the deployed model making predictions on new data?
In AI-900, “computer vision” is less about model architecture and more about mapping real scenarios (images and video) to the right Azure capability and the right kind of output. The exam repeatedly tests whether you can recognize what the customer is asking for: a label for the whole image, locations of multiple items, or text extracted from pixels. This chapter builds that decision skill by grounding you in the three core patterns—classification, detection, and OCR—then connecting them to Azure AI Vision and related choices.
You should be able to describe common vision use cases and expected outputs (labels, bounding boxes, confidence scores, extracted text), and select between prebuilt capabilities versus custom training. As you read, keep a “signal words” mindset: terms like “identify,” “detect,” “locate,” “read,” “extract,” “count,” and “verify” point strongly to a particular workload type.
Exam Tip: When a question describes the desired output, treat that as your primary clue. “Return coordinates,” “draw boxes,” or “count items” strongly indicates object detection; “return the main category” suggests classification; “extract invoice number” points to OCR/document understanding.
Practice note for Identify key computer vision use cases and outputs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map scenarios to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand detection vs classification vs OCR basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: vision scenarios and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify key computer vision use cases and outputs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map scenarios to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand detection vs classification vs OCR basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: vision scenarios and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify key computer vision use cases and outputs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map scenarios to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to distinguish the three foundational computer vision workload patterns and their outputs. Image classification answers “What is this image?” by assigning one or more labels to the entire image (often with confidence scores). A typical use case is sorting photos into categories (e.g., “cat,” “dog,” “car,” “indoor/outdoor”). The output is usually a set of tags or a single category, not locations.
Object detection answers “Where are the objects, and what are they?” It returns bounding boxes (coordinates) plus labels and confidence. This is used for counting items on a shelf, finding defects on a production line, or locating people/vehicles in a frame. The key differentiator is localization: detection includes positions.
OCR (Optical Character Recognition) answers “What text is in the image?” It extracts characters and words from pixels, often with layout information (lines, bounding regions). OCR is used for receipts, forms, IDs, and signage, and the output is text plus structure.
Exam Tip: Many candidates confuse “detect” with “classify.” In exam wording, “detect” typically implies finding multiple instances and locations. If the prompt mentions “highlight,” “box,” “coordinates,” “count,” or “track,” choose detection.
Another common trap is assuming OCR is only for documents. OCR can apply to any image containing text (street signs, containers, screenshots). If the scenario’s goal is to “read” or “extract” text, OCR is almost always the core requirement even if the image is not a PDF.
Azure’s prebuilt vision capabilities are designed to let you analyze images without training a custom model. For AI-900, focus on the conceptual mapping: you provide an image (or frame) and the service returns structured metadata. Typical outputs include tags (keywords), captions (natural language description), detected objects with bounding boxes, and extracted text via OCR.
When you see a question that asks for “a short description of what’s happening in the photo,” that aligns with image captioning. When you see “generate keywords for indexing,” that aligns with image tagging. If you see “find the location of each item,” that aligns with object detection. If you see “read serial numbers,” that aligns with OCR.
These outputs are usually returned as JSON containing fields like detected items, coordinates, confidence scores, and recognized text blocks. AI-900 does not test code syntax, but it does test whether you understand what the service returns and what you can do with it (search, indexing, automation, compliance workflows).
Exam Tip: Pay attention to whether the scenario emphasizes searchability and metadata. If the goal is to make images searchable in an app, prebuilt tagging/captioning is a strong fit. If the goal is to take an action based on a specific item’s presence and location (e.g., “if a hardhat is missing in a zone”), object detection is the better match.
Also watch for the “video” wording. AI-900 often simplifies video questions to “analyze frames” or “process snapshots.” Your answer typically still maps to the same underlying vision tasks—classification, detection, and OCR—just applied repeatedly across frames.
OCR is a high-frequency topic because it’s easy to describe in business terms: extracting values from receipts, IDs, invoices, labels, and forms. The exam expects you to know the difference between text extraction and data extraction. OCR is primarily about converting pixels to characters and words; document understanding adds structure: fields, tables, key-value pairs, and layout relationships.
When a scenario says “read the text from a sign” or “extract the product code from an image,” OCR is sufficient. When a scenario says “extract the invoice number, total, and line items,” that implies document-style understanding where layout matters. The key clue is whether the user needs just raw text, or specific fields and their meaning.
Exam Tip: Don’t overcomplicate OCR questions. If the prompt highlights “printed text,” “handwritten notes,” “receipt,” “form,” or “invoice,” your first instinct should be OCR/document extraction—not image classification.
A classic trap is picking object detection for “find the serial number label on a device.” If the goal is the text value, OCR is the core requirement; detection might help locate the label region, but the exam typically wants the service that reads characters.
Face-related scenarios introduce both capability and responsibility. AI-900 commonly tests that you can recognize privacy implications and the need for consent, transparency, and bias awareness. In practical terms, any system that detects faces, verifies identity, or infers attributes can create compliance obligations (data retention policies, access controls, and user notification).
At a conceptual level, face workflows can include detecting that a face exists in an image, locating it, and comparing two faces for similarity (verification/identification concepts). The exam focus is not on implementation detail but on when such capability is appropriate and what risks must be managed.
Exam Tip: If an answer choice mentions “consent,” “privacy,” “data minimization,” or “responsible AI,” it’s often the best companion concept to face scenarios. The exam likes to pair technical selection with governance: only collect what you need, protect biometric data, and ensure users understand how data is used.
Another common trap is assuming face analysis is always acceptable for “employee monitoring” or “customer profiling.” On AI-900, you should lean toward responsible framing: consider whether the scenario has explicit consent, a legitimate purpose, and safeguards against biased outcomes. If the scenario is identity verification, choose solutions that emphasize security and consent; if the scenario is demographic inference, consider fairness and whether it’s even appropriate.
A key exam skill is deciding when to use a prebuilt vision model versus a custom-trained model. Prebuilt capabilities are ideal when you want general understanding: common objects, generic tags, captions, and broad OCR. Custom vision is considered when your target objects are domain-specific, visually subtle, or not covered by general models (e.g., your company’s specific product SKUs, unique defect types, or specialized equipment parts).
Decision points AI-900 expects you to articulate include: the availability of labeled data, how specific the classes are, and whether the business needs consistent recognition of a narrow set of categories. If you can define categories clearly and can collect labeled images, custom training becomes viable. If you need quick results with minimal data preparation, prebuilt is usually best.
Exam Tip: “We have thousands of labeled images of our parts” is a strong hint toward custom vision. “We need to describe images for a website” or “extract printed text” usually points to prebuilt.
Watch the trap of choosing custom when the scenario doesn’t require it. AI-900 rewards the simplest fit: if the requirement is general and the output is generic tags/captions, prebuilt is correct even if custom is possible.
On exam day, you’ll often see scenario-based prompts that test service selection through the desired output. Build a fast mental checklist: (1) Is the goal text, labels, or locations? (2) Is the domain generic or specialized? (3) Is structure/layout important? (4) Are there face/privacy considerations?
For “sort images into categories,” select classification. For “count items and show where they are,” select object detection. For “read text from images,” select OCR. For “describe the scene” or “generate keywords,” select tagging/captioning. For “extract invoice totals and line items,” select document understanding (OCR plus structure).
Exam Tip: Focus on nouns and verbs in the scenario. Verbs like “locate,” “track,” “count,” and “highlight” map to detection. Verbs like “categorize,” “label,” and “classify” map to classification. Verbs like “read,” “extract,” and “transcribe” map to OCR. This reduces the chance you pick a plausible-but-wrong option.
A common trap is overfitting to the industry: “retail” does not automatically mean detection; “healthcare” does not automatically mean OCR. The exam cares about the functional output. Another trap is mixing up “analyze images” (broad) with “identify a specific custom part” (narrow). If the scenario mentions proprietary objects or a unique defect taxonomy, lean toward custom vision; otherwise, use prebuilt Azure AI Vision capabilities.
Finally, when face analysis appears, add a responsibility lens: the correct approach often includes not just the technical selection but also ensuring consent, transparency, and secure handling of biometric data. AI-900 frequently rewards answers that acknowledge these constraints because they reflect real-world Azure AI usage.
1. A retail company wants to process store shelf images and return the coordinates of every cereal box so they can count items and identify empty spaces. Which computer vision workload type best fits the requirement?
2. A company needs to categorize product photos into one of several departments (for example, "shoes," "electronics," or "home decor"). They only need a single label per image and do not need locations of objects. What is the most appropriate computer vision approach?
3. An insurance company receives photos of vehicle damage. They want to identify whether the photo contains "damage" or "no damage" and, if damage is present, highlight the damaged region with a bounding box. Which set of outputs best matches the requirement?
4. A logistics company wants to automatically read container IDs from images of shipping containers taken at a gate. The output must include the extracted alphanumeric ID string. Which workload type should you choose?
5. A manufacturer wants to verify that safety helmets are being worn in images from a factory floor camera. The solution must identify each person in the image and indicate whether a helmet is present for that person. Which capability/output is most aligned with the requirement on the AI-900 exam?
This chapter targets the AI-900 objective area that asks you to recognize natural language processing (NLP) and generative AI workloads and map them to the right Azure services. On the exam, you are rarely asked to build anything; you are asked to identify the workload type, the output you need (labels, extracted fields, summaries, answers, generated text), and then choose the best-fit Azure offering.
NLP is about understanding and transforming language: detecting sentiment, extracting entities, translating text, summarizing documents, and enabling search or question answering. Conversational AI adds speech, turn-taking, and orchestration (routing user intent to the correct tool). Generative AI goes a step further by producing new text (or other content), often guided by prompts and grounded in enterprise data.
Exam Tip: When a question mentions “extract,” “identify,” “detect,” or “classify,” think classic NLP capabilities (Azure AI Language / Speech). When it emphasizes “generate,” “draft,” “rewrite,” “create,” or “chat,” think generative AI (Azure OpenAI). Many scenarios include both; the correct answer is usually the component that performs the “hardest” step the scenario requires.
Practice note for Identify NLP tasks: sentiment, entities, summarization, translation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand conversational AI basics: speech and bots: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts and Azure OpenAI use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: NLP + GenAI service mapping and safe usage: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify NLP tasks: sentiment, entities, summarization, translation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand conversational AI basics: speech and bots: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts and Azure OpenAI use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: NLP + GenAI service mapping and safe usage: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify NLP tasks: sentiment, entities, summarization, translation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand conversational AI basics: speech and bots: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to recognize common NLP workload patterns. Start by translating the scenario into one of four patterns: classification (assign labels), extraction (pull structured data from text), Q&A (return an answer from a knowledge source), and summarization (shorten while preserving meaning). If you can name the pattern, the service choice becomes straightforward.
Classification examples: sentiment (positive/negative/neutral), topic labeling, language detection, spam detection. The output is typically a label and confidence score. Extraction examples: key phrases, named entities (people, locations, organizations), personally identifiable information (PII) detection, or custom entity extraction for domain terms (product SKUs, policy numbers). Q&A is about retrieving an answer from a curated set of documents or knowledge base rather than generating freely; the exam often signals this with phrases like “from FAQs,” “from manuals,” or “consistent answers.” Summarization is common for call transcripts, long emails, meeting notes, and legal documents.
Common trap: Confusing summarization with “generating a new report.” Summarization reduces content; generative drafting creates new content that may go beyond the source. On AI-900, the correct choice depends on whether the solution must remain faithful to the input text (summarize) or create novel output (generate).
Exam Tip: Look for constraints. “Extract order number and shipping address” is extraction. “Provide a concise version of this 3-page complaint” is summarization. “Answer questions using product documentation” is Q&A. “Assign each review a sentiment label” is classification.
Azure AI Language is the Azure service family that covers many classic NLP tasks that appear on AI-900. The exam focuses on what these capabilities do, when to use them, and what the output looks like. Three high-frequency concepts are sentiment analysis, key phrase extraction, and entity recognition.
Sentiment analysis evaluates the polarity of text (for example positive/neutral/negative) and may provide confidence scores. In scenarios like monitoring social media, customer feedback, or support tickets, sentiment helps you quantify customer mood and detect spikes in dissatisfaction. Key phrase extraction identifies important terms in a document (“late delivery,” “refund request,” “broken screen”) to support tagging, search, or analytics dashboards.
Entity recognition extracts named entities such as people, organizations, locations, dates, and sometimes categories like products or events. In business workflows, entity recognition helps convert unstructured text into fields that downstream systems can store and query. This can also include recognizing sensitive data (PII) depending on the feature set described in the question.
Common trap: Selecting a generative model to “find entities.” Generative AI can do it, but on AI-900 the test-friendly answer is typically the purpose-built NLP feature because it is more deterministic and designed for extraction at scale. Another trap is assuming “key phrases” and “entities” are the same: key phrases are salient terms; entities are items that fit defined categories.
Exam Tip: If the scenario needs structured outputs (a list of people, locations, order IDs), prefer Azure AI Language extraction/recognition. If it needs creative or open-ended writing, prefer Azure OpenAI. The exam often rewards “right tool for the job,” not “most powerful tool available.”
Speech scenarios are tested as part of NLP/conversational AI fundamentals. You should be able to distinguish speech-to-text (STT), text-to-speech (TTS), and speech translation use cases, and map them to Azure’s speech capabilities (Azure AI Speech).
Speech-to-text converts audio into text. Typical triggers in exam questions: “transcribe calls,” “convert meeting audio to text,” “caption live events,” or “create searchable transcripts.” Once you have text, you can apply downstream NLP like sentiment or summarization. Text-to-speech generates spoken audio from text, often used for accessibility, IVR systems, or reading notifications aloud. Translation appears in two flavors: translating text between languages, and translating speech in real time for multilingual conversations.
Common trap: Picking translation when the real need is transcription. If the requirement is “create a written record” or “index the content,” that is STT first. Translation is only required if the output must be in a different language. Another trap is assuming TTS is a chatbot feature; TTS is about output modality (audio), not decision-making.
Exam Tip: Watch for multi-step scenarios. “Analyze sentiment of recorded support calls” implies STT (to get text) plus sentiment analysis (to score). The best answer is often the missing capability that enables the rest of the pipeline.
Conversational AI on AI-900 is less about frameworks and more about understanding the building blocks: a bot that manages a dialogue, optional speech input/output, and orchestration that routes a user request to the right skill (FAQ lookup, ticket creation, human handoff, or a generative assistant).
A bot typically needs: (1) channels (web chat, Teams, mobile app), (2) a dialog manager to handle turns and context, (3) language understanding or intent recognition to decide what the user wants, and (4) integrations with back-end systems. Orchestration is the “traffic controller” layer—deciding whether a request should be answered by a knowledge base, a transactional workflow, or escalated to an agent.
Common trap: Treating a bot as “the AI service.” A bot is often the container experience; the AI is provided by language understanding, Q&A, or generative models behind it. Exam items may describe a “customer service chat interface” and ask what AI capability enables it—don’t pick “bot” if the actual need is intent recognition, question answering, or speech.
Exam Tip: If the scenario emphasizes “consistent answers from approved content,” think Q&A orchestration. If it emphasizes “free-form conversation” or “drafting responses,” think a generative assistant, but still note the orchestration need (routing, guardrails, and handoff).
Generative AI workloads appear on AI-900 as conceptual questions: what generative AI is, what prompts do, how to reduce hallucinations, and which Azure services support common patterns. Azure OpenAI is central: it provides access to foundation models for text generation and chat-style interactions.
Prompts are instructions plus context that steer model behavior. Good prompts specify role, task, constraints, and desired format. The exam often tests that prompts can include examples (“few-shot”) and that outputs are probabilistic—meaning the same prompt can yield different wording. Grounding is the practice of providing trusted, relevant data (for example, retrieved passages from company documentation) so the model’s answers are based on known sources rather than guessing. This is how you improve factuality and compliance in enterprise scenarios.
Embeddings are numerical representations of text that capture meaning. They power semantic search and “retrieve then generate” workflows: store embeddings for documents, find the most relevant passages for a user question, then pass those passages into a generative model. On AI-900, recognize embeddings as the enabling concept for similarity search, clustering, and retrieval-augmented generation (RAG) patterns.
Copilots describe integrated generative experiences that assist users in tools and workflows (drafting email responses, summarizing tickets, generating knowledge articles). The exam angle is usually: identify the use case (draft/summarize/chat) and the need for grounding and safety controls.
Common trap: Assuming generative AI “knows your company policies.” Unless grounded, a model only knows what was in its training data and what you provide in the prompt/context. Another trap is confusing embeddings with “generated text.” Embeddings are vectors for search and matching; they are not human-readable.
Exam Tip: When you see “semantic search,” “find similar,” “retrieve relevant documents,” or “RAG,” think embeddings + retrieval + Azure OpenAI generation. When you see “write,” “draft,” “brainstorm,” think direct generation with prompts, but still consider grounding if accuracy is required.
AI-900 questions frequently test your ability to choose between classic NLP and generative AI, and to recognize safe/appropriate usage. Your decision rule: if the required output is structured and deterministic (entities, key phrases, sentiment labels), prefer Azure AI Language. If the required output is open-ended natural language generation (drafting, rewriting, ideation, conversational responses), prefer Azure OpenAI—then add grounding when answers must be faithful to enterprise content.
Prompt basics that show up on the exam: prompts can include instructions, context, and examples; prompt quality affects output quality; and you can ask for specific formats (bullets, JSON-like structure, tone). Also remember that prompts may inadvertently include sensitive data—so safe usage includes minimizing sensitive inputs and applying data handling policies.
Safety concepts are commonly tested at a high level: mitigate harmful outputs, protect sensitive information, and ensure responsible AI. In practice, that means content filtering/moderation, restricting what data is sent to a model, logging and monitoring, and using grounding to reduce hallucinations. For user-facing assistants, orchestration and guardrails matter: route high-risk requests to human review, and ensure the assistant cites or relies on approved sources when required.
Common trap: Picking generative AI for translation or transcription. Those are classic speech/text tasks; use Speech for audio conversion and translation features for language conversion. Generative AI can paraphrase, but exam scenarios about “translate accurately” or “transcribe calls” are not asking for creative generation.
Exam Tip: In service-mapping questions, underline the verb: extract (Language), transcribe (Speech), translate (Language/Speech translation), summarize (Language or GenAI depending on wording), generate/draft/chat (Azure OpenAI). Then check for constraints like “approved sources,” “consistent answers,” and “safety,” which signal grounding and governance needs.
1. A retail company wants to automatically analyze customer reviews and return a positive, negative, or neutral label for each review. Which Azure service is the best fit?
2. A legal team needs to process thousands of contracts and extract organization names, dates, and locations into structured fields. Which capability should you use?
3. A travel website must translate user-generated hotel descriptions from French to English while keeping the meaning consistent. Which Azure service should you choose?
4. A bank wants a chatbot that can draft a personalized, natural-language explanation of a customer’s monthly spending trends based on provided transaction summaries. The output must be newly generated text. Which Azure service is the best fit?
5. Your company is building an internal assistant that uses Azure OpenAI to answer employee questions. You want to reduce the risk of the model outputting harmful or inappropriate content. What should you implement?
This chapter is your “dress rehearsal” for AI-900. The goal is not to memorize trivia, but to practice how the exam expects you to think: identify the workload, map it to the right Azure AI service family, and avoid the common traps (over-engineering, mixing legacy service names, or selecting tools that don’t match the data type or task).
You will complete two mock exam passes (Part 1 and Part 2), then run a Weak Spot Analysis to convert mistakes into repeatable rules. Finally, you’ll follow an Exam Day Checklist that reduces avoidable errors—like rushing, misreading scenario constraints, and second-guessing service boundaries.
As you work through this chapter, keep a single note: for every miss, write (1) the keyword you overlooked, (2) the service family that should have triggered, and (3) the distractor pattern that fooled you. That one page of notes is your highest ROI final review.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 tests breadth more than depth. Your time strategy should match that: move quickly, anchor on the workload type, and only then confirm the best Azure service. In the mock exam parts in this chapter, you will practice “two-pass answering.” Pass 1: answer everything you can confidently in one read. Pass 2: return to flagged items and re-check keywords (data type, real-time vs batch, labeled vs unlabeled, deterministic vs probabilistic output, and whether the requirement is to build a model or use a prebuilt capability).
Exam Tip: Treat every scenario as a classification problem of its own: “What is this question really asking me to do?” Many errors come from answering a different question than the one on screen.
Time management rules that work well for AI-900: (1) Don’t spend long debating between two similar services—flag and move; (2) Eliminate obviously wrong service families first (vision vs language vs ML vs generative); (3) For Azure AI service questions, look for the “capability noun” (translation, entity extraction, OCR, object detection, speech-to-text) and map it directly.
Common timing trap: you try to “prove” an answer by recalling product minutiae. The exam is primarily checking conceptual alignment: workload → task type → Azure service family.
Mock Exam Part 1 is designed to feel like the first half of the real test: mixed domains, mostly direct mapping, and moderate scenario detail. Your job is to practice rapid recognition across AI workloads, core ML principles, and Azure AI services for vision and language. Think in terms of “trigger words.” If you see images, scans, video frames, OCR, or detection, you should immediately consider Azure AI Vision capabilities. If you see chat, summarization, sentiment, key phrases, or entity extraction, your map should point to Azure AI Language. If you see speech-to-text, text-to-speech, or speaker recognition cues, your brain should jump to Speech.
For ML fundamentals, be ready to identify whether a task is classification (predict a category), regression (predict a number), clustering (group without labels), or anomaly detection (find unusual patterns). The exam often hides this behind business language (“predict churn,” “forecast demand,” “group customers,” “detect fraud”).
Exam Tip: When a question asks “which approach,” answer with the ML task type (classification/regression/clustering) rather than naming an algorithm, unless the options force you there.
Common distractor patterns in Part 1: choosing Azure Machine Learning for a problem that is solvable via a prebuilt Azure AI service; choosing a vision feature for a language-only input; confusing “training a custom model” with “using a pretrained endpoint.” Train custom models when the scenario demands domain-specific labels or unique categories; use prebuilt when it’s standard (OCR, general object detection, translation, sentiment).
Practice also recognizing responsible AI concepts that show up in AI-900: fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. These are often tested as “which principle is being addressed?” or “what is the risk?” Don’t overthink—match the description to the principle.
Mock Exam Part 2 increases scenario density: more constraints per question, more “choose the best fit,” and more opportunities to fall for close-but-not-quite services. This is where you must slow down enough to catch requirements like “near real-time,” “multi-language,” “offline processing,” “human review,” or “needs grounding in company documents.” Those keywords are often the whole question.
Generative AI appears more here. The exam expects foundational understanding: what a large language model (LLM) is, what “prompting” does, what embeddings are used for (semantic search), and when Retrieval-Augmented Generation (RAG) is appropriate (when answers must be based on your data rather than the model’s general knowledge). If the scenario says “use internal policies,” “use product manuals,” or “answer from enterprise documents,” the concept is almost always grounding via RAG rather than “just prompt it.”
Exam Tip: If the scenario demands citations, consistency with internal content, or reduction of hallucinations, think “RAG + embeddings + vector search,” not “bigger model” or “more temperature tuning.”
Another Part 2 trap is service boundary confusion: Azure OpenAI is for generative outputs (drafting, summarizing, chat, code); Azure AI Language is for structured NLP tasks (sentiment, entities, PII detection, classification) and conversational language understanding; Azure AI Vision is for image analysis/OCR; Azure Machine Learning is for building, training, and deploying your own models with full lifecycle control.
Also expect some “what should you do first?” sequencing items. The safest sequence is: define the business problem and success metric → confirm data availability and type → choose the approach (prebuilt vs custom) → select the Azure service → plan evaluation and monitoring. When in doubt, pick the answer that reduces risk early (data quality checks, baseline metrics, and evaluation).
This section is your Weak Spot Analysis engine. After Mock Exam Part 1 and Part 2, don’t just mark correct/incorrect—categorize the reason. Use this review framework for every missed or guessed item:
Exam Tip: Many distractors are “technically possible” but not “best answer.” The exam rewards the simplest service that meets the requirement with the least custom work.
Watch for these repeatable distractor patterns: (1) “Use Azure Machine Learning” when the scenario is a standard prebuilt AI task; (2) “Use Azure OpenAI” for tasks that are better served by deterministic extraction (entities/PII) in Azure AI Language; (3) Confusing OCR (text in images) with text analytics (text already extracted); (4) Selecting an approach (supervised vs unsupervised) opposite to whether labels exist.
When you find a recurring error, convert it into a rule you can apply under stress, such as: “If the input is an image and the output is text, it’s OCR (Vision), not Language.” These rules are your final-week advantage.
This recap ties directly to AI-900 outcomes: identify AI solution types, choose core ML approaches and Azure services, map vision and NLP scenarios to Azure AI capabilities, and explain generative AI concepts and Azure OpenAI use cases. Use it as a “mental index” during the exam.
AI workloads: Know when a problem is prediction, detection, recognition, extraction, conversation, or generation. Also know that responsible AI principles are part of “workloads” because they influence design choices and risk controls (bias checks, transparency, privacy, and human oversight).
Machine learning fundamentals on Azure: Classification vs regression vs clustering vs anomaly detection is core. Supervised learning requires labeled data; unsupervised does not. Training/validation/testing exists to prevent overfitting and to estimate real-world performance. Azure Machine Learning is the general platform for building, training, tracking, and deploying models when you need custom behavior or control over the lifecycle.
Computer vision on Azure: Map images/video to tasks like OCR (extract text), image classification (label an image), object detection (locate objects), and analysis of attributes. The key exam skill is matching the input modality (image/video) and desired output (text, tags, bounding boxes) to Azure AI Vision capabilities.
NLP and speech on Azure: Azure AI Language covers text analytics (sentiment, key phrases, entity recognition, PII) and related language understanding tasks; Speech covers speech-to-text and text-to-speech. Trap: don’t pick speech services when the scenario has only text data.
Generative AI: Understand prompts, tokens, temperature (creativity vs consistency), and why grounding matters. Azure OpenAI is used for chat, summarization, drafting, and transformation tasks. If the answer must be based on enterprise data, think embeddings + retrieval (RAG) to reduce hallucinations and improve relevance.
Your score can drop due to avoidable friction: unstable internet, notifications, rushing, or mental fatigue. Use this checklist to protect your performance and apply what you practiced in the mock exams.
Exam Tip: In the last hour before the exam, do not attempt to learn new services. Review your “rules list” from Weak Spot Analysis (e.g., modality mapping, supervised vs unsupervised cues, prebuilt vs custom decisions, RAG triggers).
Last-hour review plan (high yield): (1) Re-read your mistake patterns and the rule you wrote to prevent each one; (2) Recite the mapping: Vision ↔ images/OCR/detection, Language ↔ text analytics/entities/PII, Speech ↔ audio STT/TTS, Azure OpenAI ↔ generation and chat, Azure ML ↔ custom model lifecycle; (3) Review ML task types with one business example each; (4) Remind yourself that “best answer” usually means least complexity that satisfies constraints.
Finally, commit to reading every option fully. Many AI-900 misses happen because the first option “sounds right,” but a later option matches the scenario constraint more precisely.
1. A company wants to add a feature to its customer support app that detects whether a written message is positive, negative, or neutral. The team wants a prebuilt AI capability with minimal ML expertise. Which Azure AI service is the best fit?
2. You need to process scanned invoices (image files) and extract structured fields such as vendor name, invoice number, and total amount. You want a service designed for document understanding rather than general image labels. Which Azure service should you use?
3. A global retailer wants a chatbot that can answer questions using the company’s internal policy documents and also provide citations to the source passages. The team wants to minimize custom ML training. Which approach best matches Azure AI concepts?
4. You are reviewing practice exam results and notice you often choose a service that is too broad when a specialized service exists. Which of the following is the best example of this common certification-exam distractor pattern?
5. On exam day, you encounter a long scenario question with multiple requirements. What is the best strategy to reduce avoidable errors, based on AI-900 exam-taking best practices?