AI Certification Exam Prep — Beginner
Master AI-900 domains with clear explanations and real exam-style practice.
This Edu AI course is a focused exam-prep blueprint for the Microsoft AI-900: Azure AI Fundamentals exam—built for beginners with basic IT literacy and no prior certification experience. You’ll study exactly what the exam measures, in the same structure as the official domains, and you’ll practice with realistic, exam-style questions that train service selection, scenario reading, and elimination of distractors.
The AI-900 exam expects you to recognize common AI solution patterns and map them to the right Azure services and concepts. This course is organized into six chapters that function like a short book: orientation, four domain-focused learning chapters, and a final mock exam with review.
Chapter 1 sets you up with the exam rules, registration path, scoring expectations, and a study strategy that works for busy schedules. Chapters 2–5 each focus on one or two official exam domains, combining conceptual understanding with frequent exam-style practice milestones. Chapter 6 provides a full mock exam split into two parts plus a structured review process to diagnose weak spots and tighten your timing.
AI-900 is not about coding—it’s about recognizing scenarios and choosing the best answer. The practice in this course emphasizes:
If you’re new to Azure certifications, start by setting up your learning workflow and schedule. You can Register free to track progress and return to practice sets as you improve. Want to compare options? You can also browse all courses and pair this with other fundamentals paths.
This course is designed for individuals preparing for AI-900 who want a structured, domain-aligned plan. It’s ideal for students, career switchers, and IT professionals who need to understand Azure AI concepts—including Copilot and Azure OpenAI—at an exam-ready level.
Microsoft Certified Trainer (MCT)
Jordan Patel is a Microsoft Certified Trainer who helps beginners pass Microsoft Fundamentals exams with practical, scenario-first instruction. He has designed AI-900 exam-prep programs for teams and individuals, focusing on Azure AI services, responsible AI, and test-taking strategy.
This chapter sets your foundation for passing AI-900 with confidence and efficiency. AI-900 is a fundamentals exam, but it is not “just vocabulary.” Microsoft tests whether you can recognize real Azure AI workloads, pick the right service for the job, and apply responsible AI thinking—especially where generative AI and Copilot patterns appear in modern solutions.
Across this course you will build toward five outcomes: describe AI workloads and key considerations for choosing Azure AI solutions; explain core machine learning (ML) concepts and how Azure Machine Learning supports the ML lifecycle; identify Azure services for computer vision (classification, object detection, OCR, and responsible vision); choose Azure services for natural language processing (text analytics, translation, speech, QnA, and conversational AI); and describe generative AI concepts, Copilot patterns, and how Azure OpenAI is used securely and responsibly.
Use this chapter as your orientation and your plan. We’ll map the official objective areas to the way questions are written, clarify exam mechanics (registration, policies, question types, scoring), and then convert all of that into a 2-week and a 4-week schedule with hands-on checkpoints. Your goal is not to read more—it’s to retain more and recognize patterns under time pressure.
Practice note for Understand what AI-900 measures and how domains are weighted: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register, schedule, and take the exam (online vs test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scoring, question types, and time-management strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a 2-week and 4-week study plan with hands-on checkpoints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand what AI-900 measures and how domains are weighted: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register, schedule, and take the exam (online vs test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scoring, question types, and time-management strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a 2-week and 4-week study plan with hands-on checkpoints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand what AI-900 measures and how domains are weighted: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register, schedule, and take the exam (online vs test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 (Microsoft Azure AI Fundamentals) measures your ability to describe AI workloads and identify Azure services that implement them. Expect heavy emphasis on “service selection” and “workload fit”: given a scenario, choose the Azure offering that solves it with the least complexity and the most alignment to requirements (latency, customization, data sensitivity, and responsible AI).
Microsoft periodically adjusts domain weightings, but the exam consistently covers: (1) AI workloads and considerations, (2) fundamental ML principles on Azure, (3) computer vision workloads, (4) NLP workloads, and (5) generative AI workloads—now including Copilot-style solution patterns and Azure OpenAI security/responsible use. When you study, organize notes by these domains, not by product marketing pages.
Exam Tip: Many questions are “compare and choose” where two options sound plausible. Anchor yourself by identifying the workload type first (vision/NLP/ML/genAI), then match to the simplest Azure service that meets requirements. Over-engineering (choosing Azure Machine Learning for a straightforward prebuilt API task) is a common trap.
Another frequent trap is confusing “train a model” with “use a pretrained model.” If the scenario needs custom training on your labeled data, that pushes you toward Azure Machine Learning or custom capabilities within Azure AI services. If the scenario is standard (OCR, language detection, sentiment), prebuilt Azure AI services are usually correct.
You can take AI-900 through Pearson VUE either online (OnVUE proctoring) or at a test center. Registration is straightforward, but policy mistakes can cost you a forfeited attempt—treat logistics as part of your study plan.
Online delivery is convenient, but it is strict. You’ll be asked to photograph your testing area and present valid government-issued ID. Clear your desk completely, silence devices, and ensure a stable internet connection. Expect check-in steps and a waiting period. Test centers reduce the home-environment risk but require travel and scheduling flexibility.
Exam Tip: Choose the format that minimizes your personal risk. If you have unreliable internet, shared spaces, or frequent interruptions, schedule at a test center. A calm, predictable environment is worth more than the convenience of home testing.
Plan registration as a commitment device: pick a date that aligns with a 2-week or 4-week plan (see Section 1.5). The best candidates schedule early, then use the deadline to drive consistent practice rather than last-minute cramming.
AI-900 questions are designed to test recognition and decision-making under constraints. Expect multiple-choice and multiple-response items, plus scenario-based sets where several questions share the same context. Even when there is no formal “case study,” many items read like mini-stories: a business need, data type, and constraints, followed by “Which Azure service should you use?”
To perform well, read like an architect: highlight the nouns (image, audio, text, documents, chat) and the verbs (detect, classify, extract, translate, summarize, answer questions). Those words usually reveal the workload domain. Then look for qualifiers: “custom model,” “no-code,” “real-time,” “PII,” “on-premises data,” “responsible AI,” or “needs to cite sources.” Qualifiers are the difference between correct and almost-correct.
Exam Tip: Treat each question as a two-step problem: (1) identify the workload, (2) identify the service and the feature tier (prebuilt vs custom vs full ML). If you jump straight to product names, you’ll fall for distractors.
Time management is part of format mastery. Don’t overinvest in early questions. Mark uncertain items for review, but avoid changing answers without a clear reason—most score drops come from second-guessing.
Microsoft exams use a scaled scoring model. You don’t see a simple “X out of Y” score; you see a scaled score, and you need to meet the passing standard (commonly 700 on a 1000 scale, though the exact model is set by Microsoft). The key takeaway: not all questions necessarily contribute equally, and the difficulty distribution can vary by exam form.
Because of scaling, your strategy should be to maximize certainty on high-frequency objective areas. For AI-900, that typically means being able to confidently map scenarios to the correct Azure AI service and explain foundational ML concepts (training vs inference, evaluation, overfitting) without hesitation.
Exam Tip: Build your “must-not-miss list.” If you miss easy service-mapping items (OCR vs object detection; translation vs sentiment; Azure OpenAI vs Azure Machine Learning), you force yourself to be perfect on harder scenario questions—which is a losing trade.
Retake planning is also a professional skill. If you don’t pass, treat the score report as an objective map of weakness by domain. Do not restart from page one. Rebuild a 7–10 day remediation plan targeted to the lowest domain plus the second-lowest domain, and include hands-on reinforcement (labs) to prevent repeat confusion.
Finally, plan your exam date with a buffer. If your schedule is unpredictable, choose a date that still gives you at least a week of review flexibility without pushing you into cramming.
AI-900 rewards candidates who can retrieve information quickly and apply it to scenarios. Passive reading feels productive but produces weak recall under exam pressure. Your core strategy should combine active recall (forcing your brain to answer without looking), spaced repetition (revisiting at increasing intervals), and hands-on labs (turning abstract services into memorable workflows).
Exam Tip: If you can’t explain why an option is wrong, you don’t fully own the concept. Train with “justify the distractor”: practice naming the scenario where each wrong option would be correct.
Use two timelines depending on your availability:
Hands-on checkpoints keep your studying honest. At minimum, schedule: (1) one Azure AI Vision or OCR trial, (2) one Azure AI Language task (sentiment/translation or entity recognition), (3) one Azure Machine Learning “hello world” (even a conceptual walk-through of training/deployment), and (4) one Azure OpenAI studio experience focused on prompt design and safety controls. The goal is not depth; it is building mental anchors so exam choices feel obvious.
Microsoft Learn provides the official-aligned learning paths, terminology, and service boundaries. Use it as your source of truth for what Microsoft intends you to know—especially around updated Azure AI service names, responsible AI commitments, and generative AI guidance. Your workflow should connect Learn content to repeated retrieval practice and targeted correction.
Use a simple loop:
Exam Tip: Keep an “Azure service boundary table” as a living document. Many AI-900 errors come from boundary confusion—mixing Azure Machine Learning (custom model lifecycle) with Azure AI services (prebuilt and some customization) and with Azure OpenAI (generative models, prompt-based behavior, safety tooling).
Integrate Edu AI practice as your feedback engine: after each practice session, tag each miss with (1) domain, (2) concept type (definition vs scenario mapping vs responsible AI), and (3) the specific keyword you missed (e.g., “OCR,” “custom,” “PII,” “grounding”). Then schedule a spaced repetition review: same day quick fix, 48-hour re-test, and end-of-week mixed set. This workflow turns practice into measurable progress rather than repeated guessing.
By the end of this chapter, your objective is simple: you should know what the exam measures, how you will take it, how you will manage time and scoring expectations, and exactly how your 2-week or 4-week plan will produce retention through labs and targeted practice.
1. You are planning your AI-900 preparation and want to prioritize study time based on how Microsoft measures the skills in the exam. Which approach aligns best with the AI-900 exam orientation and objective-based preparation strategy?
2. A candidate is deciding whether to take AI-900 online or at a test center. They are concerned about avoiding exam-day disruptions and policy violations. Which action best reduces the risk of an invalidated exam attempt?
3. You are taking AI-900 and notice that some items are scenario-based and others are short knowledge checks. You want a time-management strategy that aligns with the exam’s question types and scoring approach. What should you do?
4. A company wants to build an internal Copilot-like assistant that drafts emails and summarizes meetings using company data. During AI-900 preparation, which learning outcome should you prioritize to be most exam-ready for this scenario?
5. You have two weeks to prepare for AI-900 and want a plan that matches the chapter’s recommended approach. Which study plan is most aligned with the course guidance?
This chapter targets the AI-900 “Describe AI workloads” objective and the exam’s recurring expectation: you can look at a scenario and correctly name the workload (prediction vs classification vs detection vs generation), then select the most appropriate Azure service family at a high level. You are not expected to design deep architectures, but you are expected to recognize patterns (vision, language, speech, decisioning) and apply responsible AI basics to common deployment situations.
As you study, practice “keyword-to-workload” mapping: words like forecast, estimate, probability often point to prediction/regression; which category suggests classification; find where suggests detection; create new text/images/code points to generative AI. The exam frequently tests whether you can avoid overcomplicating: many questions are solved by picking the right workload type and the right service family, not by selecting a specific model algorithm.
Exam Tip: If a prompt describes “extracting text from images,” do not label it as generative AI. That is OCR (a computer vision extraction workload). Generative AI creates new content; OCR converts existing content from one representation to another.
Practice note for Differentiate AI, ML, deep learning, and generative AI in scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match common workloads to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles to real exam-style situations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Describe AI workloads domain (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate AI, ML, deep learning, and generative AI in scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match common workloads to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles to real exam-style situations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Describe AI workloads domain (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate AI, ML, deep learning, and generative AI in scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match common workloads to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to distinguish common AI workload types based on the business question being asked. Most scenarios fit into one of four buckets: prediction, classification, detection, or generation. Your goal is to identify what the system must output.
Prediction (often regression) outputs a numeric value or probability. Typical phrasing: “predict sales next month,” “estimate time to failure,” “forecast call volume,” or “predict likelihood of churn.” The output is not a label like “fraud/not fraud,” but a number (or a probability that can later be thresholded).
Classification outputs a discrete label or category. Examples: “is this email spam?”, “which product category?”, “positive/neutral/negative sentiment,” “safe/unsafe content.” Multi-class classification chooses one of many categories; binary classification chooses between two. On the exam, classification is frequently confused with detection—watch for “where” vs “what.”
Detection identifies the presence and often the location of an item or event. In vision, object detection returns bounding boxes (where the object is in the image). In security/IoT, anomaly detection identifies unusual patterns in time-series data. Many candidates incorrectly call anomaly detection “classification,” but the scenario usually emphasizes “unusual/outlier” rather than a fixed label set.
Generation creates new content: text, code, images, summaries, or embeddings used for semantic search. Phrasing: “draft an email,” “summarize a document,” “create a job description,” “generate product images,” “write code from a prompt.” The key is novelty—content that didn’t exist before.
Exam Tip: If the system returns both a label and an explanation, it is still classification; explanation does not change the workload type. If it returns a label plus a bounding box, that’s detection.
Another exam skill is deciding whether AI is appropriate at all. Not every “automation” scenario needs machine learning or generative AI. Rules-based logic (if/then, lookup tables, deterministic workflows) is often faster, cheaper, and easier to govern—especially when requirements are stable and you can precisely define the logic.
Use rules-based logic when: the domain is deterministic; inputs are structured and limited; the organization needs full explainability; or the cost of errors is extremely high and rules can cover all cases. Example: “Reject expense claims over $5,000 without manager approval.” No model is required.
Use AI/ML when: the rules are hard to write; patterns are subtle; inputs are unstructured (text, images, audio); or the environment changes. Example: “Detect fraudulent transactions” or “classify customer emails by intent.” ML can learn patterns that are impractical to hand-code.
Use generative AI when: the goal is to produce natural language or creative output, transform text (summarize, rewrite), or enable conversational experiences that need flexible responses. But generative AI introduces extra governance needs (prompt injection, hallucinations, data leakage), so the exam may push you to pair it with responsible controls.
Common trap: Choosing generative AI for simple retrieval. If a question only needs “return the company policy paragraph,” a search or QnA approach can be more appropriate than free-form generation. A generative model may still be used, but the scenario must justify it (summarization, reasoning, conversational UX, or grounding with retrieved content).
Exam Tip: If the requirement says “must always be correct and follow strict rules,” lean rules-based or constrained workflows. If it says “unstructured input” or “too many variations,” lean AI.
At a high level, the exam focuses on recognizing Azure service families rather than memorizing every SKU. When you see vision, language, speech, or generative needs, map them to the correct family, then select the best fit within that family if the question asks.
Azure AI Services (formerly Cognitive Services) are prebuilt APIs for common workloads. This includes Vision, Speech, and Language capabilities (translation, sentiment, key phrases, named entity recognition, etc.). If a scenario describes “call an API to analyze an image or text” with minimal training, Azure AI Services is a strong default.
Azure Machine Learning supports the full ML lifecycle: data prep, training, tracking, deployment, and monitoring. Choose it when the scenario emphasizes building a custom model, training on the organization’s data, or MLOps (pipelines, model registry, endpoints). This is a major differentiator: prebuilt AI Services vs custom ML in Azure ML.
Azure OpenAI Service is used for generative AI with large language models and image generation models (depending on availability). Scenarios: chat assistants, summarization, content generation, code generation, embeddings for semantic search. The exam often expects you to mention secure usage: authentication, network controls, and responsible filters.
Azure AI Search (often paired with Azure OpenAI) supports indexing and retrieval over enterprise content. If the scenario mentions “search across documents,” “semantic ranking,” or “ground responses in internal PDFs,” think AI Search as the retrieval component—even if generation is done by Azure OpenAI.
Common trap: Picking Azure Machine Learning for simple OCR or translation. If no custom training is needed and a prebuilt API exists, the exam usually expects Azure AI Services. Conversely, if the scenario explicitly says “train a model with your labeled images,” Azure Machine Learning (and/or custom vision capabilities) becomes more relevant.
Exam Tip: Look for verbs: “train,” “experiment,” “deploy model endpoint,” “monitor drift” → Azure Machine Learning. “Analyze image/text,” “extract,” “translate,” “transcribe” → Azure AI Services. “Generate/summarize/chat/embeddings” → Azure OpenAI.
AI-900 includes responsible AI at a conceptual level: you must recognize risks and the appropriate principle to apply. Three commonly tested principles are fairness, reliability & safety, and privacy & security. The exam does not expect legal expertise; it expects correct identification of the concern and a sensible mitigation direction.
Fairness means the system should not produce biased outcomes for groups defined by sensitive attributes (for example, gender, age, ethnicity) or proxies. Exam scenarios often describe different error rates across groups (higher false rejects for one group). The correct response is to evaluate bias, improve representative data, and monitor model performance by subgroup.
Reliability and safety means the system behaves consistently under expected conditions and fails safely. In generative AI, this includes reducing harmful outputs and ensuring grounded responses. In traditional ML, it includes robustness to data changes and careful threshold selection (e.g., false positives vs false negatives in medical screening).
Privacy and security covers protecting personal and confidential data: least privilege access, encryption, data minimization, and preventing unintended disclosure. For generative AI, privacy concerns include prompts containing sensitive data and outputs leaking confidential information. For vision, privacy concerns include facial recognition use and storing images longer than necessary.
Common trap: Confusing privacy with transparency. If the issue is “users don’t know how decisions are made,” that’s transparency. If the issue is “data could expose personal information,” that’s privacy.
Exam Tip: When a scenario mentions protected classes or unequal treatment, answer with fairness. When it mentions outages, unsafe behavior, or inconsistent results, answer with reliability/safety. When it mentions PII, consent, retention, or access controls, answer with privacy/security.
The remaining responsible AI principles frequently appear in scenario form: transparency, inclusiveness, and accountability. The exam wants you to connect these principles to practical actions in Azure-based solutions.
Transparency means stakeholders understand what the system can and cannot do. Practically: communicate that outputs may be probabilistic; provide explanations where possible; document data sources and limitations; and label AI-generated content when appropriate. In Azure, transparency is often addressed through clear user experience design (disclosures), model documentation, and logging for traceability.
Inclusiveness means the solution is accessible and usable by people with diverse abilities and backgrounds. Scenario cues: speech systems struggling with accents, captions needed for hearing-impaired users, vision models missing certain lighting/skin tones, or a chatbot that fails non-native language phrasing. The mitigation is broader testing, representative datasets, and fallback user experiences (human handoff, alternative input methods).
Accountability means humans remain responsible for outcomes. You typically implement governance: role assignments, review processes, audit logs, and escalation paths. For generative AI copilots, accountability often means “human in the loop” approval before actions are taken (sending emails, updating records) and monitoring for misuse.
Common trap: Treating accountability as “the model is accurate.” Accuracy is performance; accountability is ownership and oversight. If the scenario mentions “who is responsible” or “approval/audit,” it’s accountability.
Exam Tip: If the prompt says “users rely on outputs as facts,” transparency is the missing principle—add disclosures, citations/grounding, and guidance on verification. If it says “the system must work for all users,” inclusiveness. If it says “ensure oversight and auditability,” accountability.
This domain is often scored through short scenarios where two answers seem plausible. Your strategy is to (1) identify the workload type, (2) select the Azure family that best matches the level of customization, and (3) apply responsible AI if the scenario includes risk cues.
First, underline the output: number/probability (prediction), label (classification), location + label (detection), extracted text (OCR extraction), or new content (generation). Then check whether training is required. If the scenario says “use a prebuilt API,” “no data science team,” or “minimal setup,” choose Azure AI Services. If it says “train on our labeled dataset,” “track experiments,” or “deploy our own model,” choose Azure Machine Learning. If it says “draft, summarize, chat, or generate,” choose Azure OpenAI (often with retrieval patterns using Azure AI Search when internal documents are involved).
Next, scan for responsible AI triggers. Mentions of unequal impact across groups should immediately shift you to fairness considerations. Mentions of sensitive data, customer records, or regulations should shift you to privacy/security controls (access, encryption, data minimization). Mentions of unsafe behavior, unpredictable outputs, or critical decisions should trigger reliability/safety plus human oversight.
Exam Tip: Many wrong answers are “near misses” (right service, wrong workload). For example, selecting a generative service when the task is sentiment analysis (classification) is a common distractor. If the scenario is about “analyze,” “extract,” or “detect,” default to non-generative workloads unless it explicitly asks to create new content.
Finally, remember the exam’s level: it tests recognition, not implementation detail. You rarely need to name a specific model architecture. If you can confidently say “this is classification using a prebuilt language API” or “this is generation using Azure OpenAI with governance controls,” you are answering at the expected AI-900 depth.
1. A retail company wants to estimate next month’s sales for each store using historical sales, promotions, and holidays. Which AI workload is being described?
2. You need to extract printed text from scanned invoices and store it as searchable text. Which Azure AI service family is the best fit at a high level?
3. A manufacturer wants to identify whether each product coming off an assembly line is "defective" or "not defective" based on sensor readings. What workload is this?
4. A city wants an AI solution that locates pedestrians and bicycles in traffic camera footage by drawing bounding boxes around them. Which workload is being described?
5. A bank deploys an AI model to help approve loans. After release, it discovers approval rates differ significantly across demographic groups. Which Responsible AI principle is most directly being violated?
This chapter maps directly to the AI-900 objective area Fundamental principles of ML on Azure: you must explain core machine learning concepts (features, labels, training vs inference, overfitting/underfitting) and connect them to how Azure Machine Learning (Azure ML) supports the ML lifecycle. The exam is not asking you to derive math; it is asking you to recognize ML workload types, choose the right approach, and identify Azure ML components that enable training and deployment.
Expect scenario-style questions: “You have historical labeled data…,” “You don’t have labels…,” “Your model performs well in training but poorly in production…,” or “You need repeatable training and deployment….” Your job is to translate those phrases into the correct learning type and the correct Azure ML capability.
Exam Tip: On AI-900, the fastest path to correct answers is vocabulary matching. If you see label, think supervised. If you see group similar, think clustering. If you see deploy for predictions, think endpoint/inference. If you see repeatable workflow, think pipeline.
The sections below walk through supervised vs unsupervised learning, training/validation/testing and generalization, and a practical tour of Azure ML concepts (workspace, compute, pipelines, endpoints). The chapter ends with a practice set section (no questions embedded here) to guide what you should be able to do under exam time pressure.
Practice note for Understand supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain training, validation, testing, and overfitting/underfitting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Walk through the Azure Machine Learning workspace and ML lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML principles on Azure (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain training, validation, testing, and overfitting/underfitting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Walk through the Azure Machine Learning workspace and ML lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML principles on Azure (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Machine learning problems on the AI-900 exam are described using a small set of core terms. Features are the input variables (columns) used to make a prediction: age, temperature, transaction amount, image pixels, or text embeddings. A label (also called target) is what you want to predict: churn yes/no, house price, defect category, or fraud vs not fraud. When you run a trained model on new data to produce outputs, that is inference.
The exam frequently checks whether you can pick the right evaluation metric for the objective. For classification, you will see accuracy, precision, recall, and F1 score. Accuracy is overall correctness but can be misleading for imbalanced classes (for example, 99% non-fraud). Precision answers “when the model predicts positive, how often is it right?” Recall answers “of all actual positives, how many did we catch?” For regression, look for MAE (mean absolute error), MSE/RMSE, and R-squared—metrics that measure how close numeric predictions are to actual values.
Exam Tip: If the scenario emphasizes “minimize false alarms” (false positives), lean toward improving precision. If it emphasizes “don’t miss any” (false negatives), lean toward improving recall. These words often appear verbatim in item stems.
Common trap: mixing up the “output” of a model with the “label.” Labels are known during training; outputs are produced during inference. Another trap is assuming “accuracy” is always the best metric. If the question mentions rare events, skewed data, or compliance/safety risks, accuracy alone is often not the best choice.
Supervised learning means you have labeled examples and you want the model to learn a mapping from features to labels. AI-900 expects you to identify whether the label is numeric (regression) or categorical (classification), then match the scenario to a typical algorithm family. You are not tested on implementation details, but you are tested on recognizing what fits.
Regression predicts a number: forecasting demand, estimating delivery time, predicting energy usage, or valuing a property. Common algorithms include linear regression and decision tree–based regression. Classification predicts a category: spam vs not spam, defect type A/B/C, customer segment label, or disease positive/negative. Common algorithms include logistic regression, decision trees, and support vector machines; in modern practice you may also see boosted trees and neural networks, but the exam tends to stay conceptual.
Exam Tip: If the prompt includes “probability of …” and the outcome is yes/no, it is still classification (often logistic regression conceptually). Probability outputs do not automatically make it regression.
Another frequent exam pattern is mapping the business goal to the correct learning type. “Predict whether a customer will churn” is classification; “predict how many days until churn” is regression. Also watch the trap where a scenario includes text or images: the data type can be unstructured, but the learning task is still classification/regression depending on the label.
Finally, supervised learning depends on data quality: label accuracy, feature relevance, and representativeness. If the question hints that labels are missing or unreliable, supervised learning may be inappropriate—or you may need more data preparation and validation controls before training.
Unsupervised learning is used when you do not have labels and you want to discover structure in the data. On AI-900, the two most testable unsupervised patterns are clustering and anomaly detection. Clustering groups similar items together—customer segmentation, grouping documents by topic, or organizing product catalogs. A classic algorithm is k-means, but you mainly need to recognize the “group similar without labels” intent.
Anomaly detection looks for unusual behavior compared to a baseline: detecting suspicious transactions, equipment sensor spikes, network intrusions, or rare manufacturing defects. The exam will often describe anomalies as “outliers,” “unexpected patterns,” or “deviations from normal.” Importantly, anomaly detection can be unsupervised (no labeled fraud/normal) or semi-supervised (trained mostly on normal data). Your clue is whether labeled examples of the abnormal class exist.
Exam Tip: If the scenario asks you to “create segments” or “group by similarity,” do not choose classification just because the output looks like a category. Classification requires known labeled categories; clustering creates groups based on the data.
Common trap: confusing anomaly detection with binary classification. If the question explicitly says you have labeled examples of fraud vs not fraud, classification is viable. If it says “we don’t know what fraud looks like yet” or “no historical labels,” anomaly detection is the better conceptual fit.
Unsupervised methods are also used for exploratory analysis before supervised training, such as discovering new customer segments that later become labels. The exam may test this indirectly by asking which approach helps “understand” data rather than “predict” an outcome.
The ML lifecycle is a frequent AI-900 theme: prepare data, train a model, evaluate it, and deploy it for inference. The exam checks whether you can distinguish training, validation, and testing and recognize overfitting and underfitting symptoms.
Data preparation includes cleaning (handling missing values), transforming (encoding categories), normalizing/scaling when appropriate, and splitting data into sets. Training fits model parameters using the training set. Validation is used to tune hyperparameters and select a model configuration. Testing is a final, unbiased evaluation on unseen data to estimate real-world performance.
Overfitting occurs when a model learns training noise: high training performance, low validation/test performance. Underfitting occurs when the model is too simple or not trained enough: poor performance on both training and validation/test. Questions often describe these patterns in plain language (for example, “works great on training data but fails in production”).
Exam Tip: Watch for data leakage, a common trap: if the same customer appears in both training and test, or if a feature indirectly contains the label (like “closed_date” when predicting “will close”), the model will look unrealistically good. The exam may not use the term “leakage,” but it will describe suspiciously perfect metrics.
Another trap is mixing up validation and testing. Validation is for tuning during development; testing is the final check. If a question asks which set you use to “choose hyperparameters,” that is validation. If it asks which set provides the “final estimate” of performance, that is test.
Azure Machine Learning (Azure ML) is Azure’s primary service for managing the ML lifecycle at scale. For AI-900, you should be able to describe what an Azure ML workspace is and identify key building blocks: compute, data, experiments, pipelines, and endpoints. The exam typically stays at the “what is it for?” level, not SDK syntax.
A workspace is the top-level container that organizes assets (datasets, models, jobs, endpoints) and ties into Azure security and governance. Compute is where work runs. You will see compute instances (interactive dev) and compute clusters (scalable training). The practical exam cue: if the question needs scale-out training, choose a cluster; if it needs a personal dev machine in the cloud, choose an instance.
Pipelines represent repeatable workflows: data prep steps, training steps, evaluation, and registration. They matter for consistency and MLOps-style automation. Endpoints are for deployment/inference. A real-time endpoint supports low-latency online predictions; batch inference is for scoring large datasets asynchronously. The exam may simply say “deploy the model so apps can call it,” which points to endpoints.
Exam Tip: Separate “training” words from “serving” words. Train, tune, experiment → compute/pipelines. Deploy, consume, predict, call from an app → endpoint/inference.
Common trap: selecting Azure AI services (prebuilt) when the prompt clearly needs custom model training. If the scenario says you have your own labeled dataset and want to train and deploy a custom model, Azure ML is the expected direction.
This section is your checklist for exam readiness on “Fundamental principles of ML on Azure.” Even when questions are wrapped in business context, they usually reduce to a few decisions: supervised vs unsupervised, classification vs regression, which dataset split is being discussed, what metric matters, and which Azure ML component enables the requirement.
When you do an exam-style practice set, force yourself to underline the clue words in the stem: “labeled,” “predict a number,” “group similar,” “outlier,” “hyperparameter tuning,” “deploy,” “low latency,” “batch scoring,” “repeatable workflow.” Then map each clue to a concept: learning type, metric, or Azure ML asset.
Exam Tip: If two answers both sound “ML-ish,” pick the one that matches the operational requirement. Example patterns: “repeatable and auditable training” points to pipelines; “scale training” points to clusters; “app calls for predictions” points to endpoints.
Common traps to watch during practice: (1) choosing accuracy for imbalanced problems, (2) treating clustering results as “labels” and calling it classification, (3) using the test set to tune, and (4) confusing training infrastructure (compute) with deployment infrastructure (endpoint). If you can consistently avoid these, you are aligned with what AI-900 typically tests in this chapter’s objective area.
1. A retail company has two years of transaction history where each record includes customer attributes (age, region, average basket value) and a known outcome indicating whether the customer churned. The company wants to predict churn for current customers. Which machine learning approach should you use?
2. You train a model in Azure Machine Learning and it achieves very high accuracy on the training dataset but performs significantly worse on the test dataset. What is the most likely issue?
3. A data science team wants to tune hyperparameters for a model in Azure Machine Learning. They need a dataset split that they can use repeatedly during model selection without using the final unbiased dataset reserved for evaluation. Which dataset split should they primarily use for this purpose?
4. A company is building an ML solution on Azure and wants a repeatable process that prepares data, trains the model, and registers the resulting model in a consistent way each time the workflow runs. Which Azure Machine Learning capability best supports this requirement?
5. A manufacturer has sensor readings from machines but no labels indicating failure types. They want to group machines into sets with similar operating behavior to identify patterns. Which technique should they use?
This chapter maps directly to the AI-900 objective area Computer vision workloads on Azure: identifying vision tasks, choosing the right Azure service for image classification, object detection, OCR, and responsible vision solutions. On the exam, you are rarely asked to write code—you are tested on whether you can recognize the workload (what problem is being solved) and select the correct Azure capability or service family for that workload.
Expect scenario-style questions: “A retail company wants to…” or “A healthcare provider needs to…” Your job is to translate the scenario into a vision task (classification vs detection vs OCR vs document understanding), then choose the best-fit Azure service. The most common trap is picking a tool based on a keyword (“image,” “document,” “text”) rather than the actual output required (labels, bounding boxes, extracted text, or structured fields).
We’ll walk through core vision task types, Azure AI Vision capabilities, and how OCR differs from document processing. We’ll also cover security/privacy and responsible vision patterns—an increasingly tested area because it affects deployment decisions, not just model accuracy.
Practice note for Identify vision tasks and pick the right Azure service per scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand OCR, image analysis, and document processing basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Review security, privacy, and responsible vision considerations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Computer vision workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify vision tasks and pick the right Azure service per scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand OCR, image analysis, and document processing basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Review security, privacy, and responsible vision considerations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Computer vision workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify vision tasks and pick the right Azure service per scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand OCR, image analysis, and document processing basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to distinguish the “shape” of the output for common computer vision tasks. This is the fastest way to identify the correct answer in a scenario question.
Image classification assigns one (or more) labels to an entire image. The output is a label and a confidence score, such as “dog: 0.98” or “damaged-package: 0.87.” Use classification when the question is about deciding what the image is overall.
Object detection identifies and localizes objects in an image. The output includes labels plus bounding boxes (coordinates). Use detection when the question asks where items are, how many there are, or if you must draw boxes around items (people, cars, defects on a surface).
Segmentation is more granular than detection: it assigns a class to each pixel (or produces a mask) to outline objects precisely. Segmentation is used when bounding boxes are not precise enough—for example, measuring the area of a tumor region, calculating the amount of spilled liquid, or separating foreground/background.
In this course’s “pick the right Azure service per scenario” lesson, you’ll repeatedly apply this translation step: first identify the vision task, then map to the Azure capability that produces the required output.
Azure vision questions typically revolve around Azure AI Vision capabilities that provide prebuilt analysis of images and videos. The exam focus is not on the API surface, but on what you can extract: tags, captions, objects, and text.
Image analysis refers to general-purpose understanding of image content. You may see features like generating tags (keywords), creating a caption/description, detecting common objects, or analyzing image properties. These are best when you don’t want to train a custom model and the goal is broad metadata enrichment (for search, accessibility, or content organization).
OCR (optical character recognition) extracts printed or handwritten text from images. On AI-900, OCR is frequently a distinct requirement: “extract text from a photo,” “read serial numbers,” “digitize scanned pages,” or “capture text from a storefront sign.” OCR output is text plus layout-related hints (lines/words), depending on the capability.
Exam Tip: If the question asks for “extract text” or “read text,” OCR is the primary feature. If it asks for “extract invoice number” or “total amount,” that’s a document processing workload (Section 4.3), even though OCR will likely be involved behind the scenes.
This section ties to the lesson “Understand OCR, image analysis, and document processing basics.” For the exam, practice separating “free-form text extraction” (OCR) from “field extraction into named properties” (document intelligence).
Document processing is a specialized subset of vision workloads where the goal is not merely to read text, but to convert documents into structured data. AI-900 typically frames this as forms processing, invoice/receipt extraction, and ID document scenarios.
In a forms scenario, the user might submit scanned forms with checkboxes, typed entries, or handwritten fields. The expected output is a set of key-value pairs (for example: “EmployeeName=…”, “StartDate=…”) and often table extraction. In receipts, you’ll commonly see fields like merchant name, date, subtotal, tax, and total. In IDs (where allowed by policy and region), the scenario may ask to extract fields such as name, date of birth, or document number.
From an architecture perspective, document processing often includes validation steps (for example, cross-check totals, verify required fields) and can feed downstream systems like CRM/ERP. On AI-900, you won’t be tested on integration mechanics, but you may be tested on selecting the correct service category for “automate data entry from documents.”
A major decision point in Azure vision workloads is whether to use prebuilt capabilities or build a custom model. The exam tests your ability to choose customization when the scenario contains domain-specific requirements that prebuilt models can’t reliably satisfy.
Prebuilt solutions are ideal when the problem matches common patterns: general tagging, captions, common object detection, or generic OCR/document extraction. They are faster to implement, require little to no training data, and are typically cost-effective for proof-of-concepts.
Custom vision solutions matter when you must recognize company-specific items (custom product SKUs, proprietary components, specialized medical imagery, unusual defect types) or when you need to enforce your own label taxonomy. Customization usually implies you will supply labeled images and iterate to achieve required accuracy.
In “Identify vision tasks and pick the right Azure service per scenario,” your fastest method is: (1) identify task type, (2) decide prebuilt vs custom based on domain specificity and labeling needs, (3) select the service family aligned to that output.
AI-900 increasingly emphasizes responsible AI considerations. For vision workloads, this commonly shows up as questions about privacy, PII, consent, and human review for sensitive decisions. You should assume that images can contain personal data (faces, license plates, ID cards, medical information) even if the scenario does not explicitly call it out.
PII and security considerations: Minimize data collection, store images securely, apply encryption, and restrict access via least privilege. In many architectures, you should avoid storing raw images unless necessary; store derived results (extracted fields, labels) when that meets the requirement. Redaction or blurring is a common mitigation when images are used for analytics but contain sensitive regions.
Bias and fairness risk: Vision models can underperform across lighting conditions, skin tones, camera types, or geographic contexts. The exam won’t require a deep statistical treatment, but it will test whether you recognize that you must evaluate model performance across representative samples and monitor drift over time.
Human-in-the-loop (HITL) patterns: For high-impact outcomes (identity verification, safety incidents, compliance flags), build a review workflow where uncertain predictions or low-confidence results are routed to a human. Combine confidence thresholds with audit logs to support accountability.
This section aligns to the lesson “Review security, privacy, and responsible vision considerations.” If an answer choice mentions governance actions (auditability, access controls, human review, minimizing data), it is often the best-fit in responsible AI questions.
This chapter’s practice set (provided separately) is designed to mirror AI-900’s scenario style without requiring implementation details. To score well, focus on the decision framework the exam expects: identify the workload, match the output type, then pick the simplest Azure service that satisfies requirements.
Step-by-step method for exam questions:
Exam Tip: When two answers both “sound plausible,” choose the one that matches the required output most precisely. AI-900 favors correct workload mapping over broader, catch-all descriptions.
Common traps to watch for in practice: (1) Selecting OCR when the scenario asks for named fields like totals and dates; (2) selecting general image tagging when the scenario needs localization/counting; (3) selecting custom training when the scenario explicitly wants a quick deployment using prebuilt capabilities; (4) ignoring privacy/HITL requirements in scenarios involving IDs or individuals.
Use the practice set to drill pattern recognition: under timed conditions, you should be able to classify the scenario into one of the few core categories in under 15–20 seconds, then spend the remaining time eliminating distractors.
1. A retail company wants to analyze in-store camera images to identify whether a shelf is empty and highlight the empty shelf area in the image. Which computer vision task and Azure service best fit this requirement?
2. A healthcare provider scans patient intake forms that contain printed text and handwritten entries. They need the text extracted from the scans, but they do not need to map values to specific fields. Which Azure capability should you use?
3. An insurance company receives thousands of claim documents. They must extract structured information such as policy number, claimant name, and claim amount from specific parts of each document. Which Azure service is the best fit?
4. A company wants to classify product images as 'acceptable' or 'damaged' before listing them online. The images show products on a plain background, and no bounding boxes are required. Which approach should you choose?
5. A financial services company plans to process customer-submitted ID photos to extract text and verify details. They want to reduce privacy risk and support responsible AI practices. Which design choice best aligns with security and privacy considerations for vision workloads on Azure?
This chapter maps directly to AI-900 objectives around NLP workloads on Azure and Generative AI workloads on Azure. The exam expects you to recognize common text, translation, speech, and conversational scenarios and choose the correct Azure service (often by spotting keywords like “sentiment,” “entity extraction,” “translation,” “speech-to-text,” “chatbot,” or “grounding with enterprise data”). You are not expected to design a production architecture, but you are expected to pick the right managed capability and describe what it does at a high level.
We’ll start with how NLP workloads are categorized (classification, extraction, summarization, and translation), then connect those workloads to Azure AI Language and Speech services. Finally, we’ll shift to generative AI: what LLMs do, why tokens/embeddings matter, and how Azure OpenAI + Copilot patterns use grounding/RAG and safety controls. Throughout, focus on how the exam phrases requirements and how to eliminate distractors that mention the wrong service family or an overly complex option.
Exam Tip: On AI-900, “choose the service” questions are usually won by matching the task verb (analyze, extract, translate, transcribe, answer, generate) to the correct Azure AI capability and ignoring unrelated platform components (Kubernetes, networking, data lakes) unless the prompt explicitly asks about them.
Practice note for Choose Azure services for sentiment, key phrases, NER, translation, and speech: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Design conversational solutions: question answering and bots at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts and Copilot patterns for business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Use Azure OpenAI concepts: prompts, grounding, safety, and deployment basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: NLP + Generative AI workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose Azure services for sentiment, key phrases, NER, translation, and speech: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Design conversational solutions: question answering and bots at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts and Copilot patterns for business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Use Azure OpenAI concepts: prompts, grounding, safety, and deployment basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
NLP workloads on the AI-900 exam are typically framed by what you want to do with text: classification (assign labels), extraction (pull structured info out), summarization (shorten content while preserving meaning), and translation (convert languages). The exam often gives a business story (support tickets, reviews, emails, call transcripts) and asks which Azure service or feature to use.
Classification includes sentiment analysis (positive/negative/neutral), topic labeling, or categorizing documents by type. ExtractionSummarizationTranslation
On Azure, these tasks generally map to Azure AI services rather than building and training a model from scratch. For AI-900, treat managed APIs as the default answer unless the question explicitly says you must train a custom model or you have highly specialized labels that require custom training.
Common trap: Confusing “summarize” as a classic NLP feature vs. generative summarization. If the question emphasizes “natural-sounding, rewritten summary” or “draft executive brief,” that points to generative AI (Azure OpenAI). If it emphasizes “extract key sentences” or “extract highlights,” it can point to language summarization features. Use the wording: “generate” and “draft” are strong generative cues.
Exam Tip: When you see “extract entities,” “key phrases,” or “sentiment,” think Azure AI Language first. When you see “translate,” think Translator. When you see “transcribe,” think Speech. When you see “write,” “compose,” or “create,” think Azure OpenAI.
Azure AI Language is the exam’s core service family for text analysis. At a fundamentals level, you should be able to describe what the common features do and when you would choose them: sentiment analysis, named entity recognition (NER), and key phrase extraction. These features help transform unstructured text into structured signals for dashboards, routing, and automation.
Sentiment analysis evaluates the emotional tone of text (for example, customer reviews). In exam scenarios, look for language like “determine whether feedback is positive or negative” or “measure customer satisfaction from surveys.” NER identifies real-world entities (people, organizations, locations, dates, product names) and is often used for indexing, compliance, and search. Key phrase extraction pulls the most important terms from a document; it’s commonly used for tagging support tickets or summarizing themes at scale.
The exam also expects you to understand that these are prebuilt capabilities—fast to adopt, minimal ML expertise required. If a question says the organization needs a custom taxonomy (for example, domain-specific categories), then “custom text classification” can be a better fit than generic sentiment or key phrases.
Common trap: Confusing NER with key phrase extraction. NER returns entities with types (Person, Location, Organization), while key phrases returns important terms without entity typing. If the question requires “identify names and addresses” or “find company names,” that’s NER (and potentially PII detection). If it requires “identify main topics” or “auto-tag keywords,” that’s key phrases.
How to identify the correct answer: underline the noun being extracted. If it’s “entities” (names, places), choose NER. If it’s “keywords,” choose key phrases. If it’s “tone/opinion,” choose sentiment.
Exam Tip: If a scenario mentions compliance, privacy, or redaction (for example, “remove credit card numbers before storing text”), think PII detection within language capabilities rather than sentiment/NER alone—AI-900 often tests your ability to spot the governance requirement hidden in the story.
Speech and conversational AI questions test whether you can distinguish between speech processing (turn audio into text or text into audio) and conversation orchestration (managing dialogues, intents, and responses). On Azure, speech-to-text and text-to-speech are classic Speech service capabilities, while bots and Q&A experiences use conversational building blocks.
Speech-to-text appears when the prompt mentions “transcribe calls,” “convert meeting audio to text,” or “create captions.” Text-to-speech
For conversational solutions, the exam frequently contrasts two approaches: (1) question answering over a knowledge base (FAQ-style) and (2) bots that manage multi-turn interactions. A Q&A solution is appropriate when users ask factual questions and expect consistent answers sourced from curated content. A bot becomes relevant when the conversation requires state, branching logic, or integration actions (e.g., “reset my password,” “check my order status”).
Common trap: Treating speech-to-text as the same as “language understanding.” Speech-to-text only produces text; you may still need language analysis or a conversational layer to interpret intent or answer questions. Another trap is assuming every chatbot must be generative—many exam questions still expect a deterministic Q&A system for predictable support answers.
Exam Tip: If the scenario says “users ask questions based on internal documents/FAQs,” pick a question answering capability. If it says “users need a conversational interface that can perform tasks,” pick a bot framework approach. If it says “audio,” always consider Speech first before language or OpenAI.
Generative AI questions on AI-900 validate that you understand what an LLM is and what it is not. An LLM (large language model) predicts the next token in a sequence, enabling it to generate text, summarize, rewrite, classify, and answer in natural language. The exam focuses on foundational terms: tokens, embeddings, and common prompt patterns.
Tokens are chunks of text used by the model (not necessarily words). Token limits affect how much input plus output you can fit into a single request, which matters when the scenario mentions “long documents” or “conversation history.” Embeddings are vector representations of meaning; they enable semantic search and similarity matching. When the prompt says “find the most relevant passages” or “search by meaning, not keywords,” embeddings are the concept being tested.
Prompt patterns show up as “how do you instruct the model?” Common patterns include: providing clear instructions, giving examples (few-shot), setting a role, specifying output format, and adding constraints (tone, length, citations). For exam purposes, you should recognize that better prompts reduce ambiguity and help produce consistent outputs, but prompts are not a security boundary.
Common trap: Overstating determinism. LLM outputs can vary; even with strong prompts, generative results are probabilistic. If the scenario requires strict, repeatable outputs or guaranteed factuality, the correct design often includes grounding with trusted data and/or deterministic components rather than “just prompt the model.”
Exam Tip: Watch for phrasing like “semantic similarity,” “vector,” “closest match,” or “retrieve relevant content”—that is your clue to embeddings and retrieval patterns, not classic keyword search and not sentiment/NER.
Azure OpenAI is Microsoft’s managed offering for using OpenAI models with Azure enterprise controls. The exam wants you to understand secure and responsible usage patterns: deployment basics, prompting, grounding, RAG (retrieval augmented generation), and safety filters. You typically won’t be asked for API syntax; you will be asked to choose the right approach and describe what it accomplishes.
Deployments in Azure OpenAI represent configured model endpoints in your Azure subscription. Exam questions may use the word “deploy” to mean “make a model available for inference with specific settings.” This is different from deploying an app service or VM.
Grounding means anchoring responses in trusted, relevant information—often your organization’s data—so the model is less likely to hallucinate and more likely to answer with correct context. The most common grounding pattern is RAG: retrieve relevant passages (often using embeddings + vector search) and provide them to the model as context for generation. In Copilot-style business scenarios (HR policy assistant, IT helpdesk, contract Q&A), RAG is the typical pattern because it keeps the model up to date without retraining.
Safety and responsible AI show up in requirements like “prevent harmful content,” “filter hate/violence/sexual content,” “reduce jailbreak risk,” or “protect sensitive data.” Azure OpenAI includes content filtering and monitoring options, but the exam expects you to know the high-level idea: you apply safety controls and you design to minimize exposure of sensitive information. Also recognize that grounding improves factuality but does not automatically guarantee compliance; you still need access controls and data governance.
Common trap: Thinking RAG “trains the model on your data.” RAG does not fine-tune by default; it retrieves and injects context at runtime. If a question explicitly asks to “adapt the model to a domain with new behavior and consistent tone,” that leans toward fine-tuning (though AI-900 usually emphasizes the concept more than the mechanics).
Exam Tip: When you see “must answer using our documents,” “must cite sources,” or “reduce hallucinations,” pick grounding/RAG. When you see “block unsafe content,” pick safety filters/content moderation controls. When you see “enterprise secure access,” emphasize Azure-managed deployments and governance rather than public, unmanaged endpoints.
This chapter’s practice set should train your “service selection reflex,” which is exactly what AI-900 tests. The exam questions are rarely about deep implementation; they are about matching a requirement to the correct workload and Azure service family while avoiding distractors that sound plausible but don’t meet the requirement.
When you review practice items, use a consistent elimination method:
Also watch how questions try to bait you with overlapping terminology. “Summarize” could mean extractive highlights (language features) or generative rewriting (Azure OpenAI). “Chatbot” could mean a bot with orchestrated dialogs, or a Q&A assistant, or a generative assistant with grounding. Your job is to match the business need: consistent factual answers (Q&A), task completion (bot), or drafting and flexible language (generative).
Exam Tip: If the scenario mentions “internal knowledge base,” “company policies,” “product manuals,” or “SharePoint content,” assume grounding is expected. If it also mentions “reduce hallucinations” or “cite sources,” that is a strong RAG signal. If it mentions “detect sentiment,” “extract entities,” or “key phrases,” that is a strong Azure AI Language signal.
Finally, remember that AI-900 often tests responsible AI posture implicitly. If two answers both seem to work, choose the one that includes safety, governance, and using managed services appropriately—those clues are frequently the tie-breaker.
1. A retail company wants to analyze customer reviews to determine whether each review is positive, negative, neutral, or mixed. Which Azure service should you use?
2. You need to extract people, organizations, and locations from support tickets stored as text. Which Azure capability best fits this requirement?
3. A travel app must translate short user messages from French to English in real time. Which Azure service should you choose?
4. A company wants to build a customer-facing chatbot that answers questions using content from their internal policy documents. They want responses grounded in that data rather than purely generated from the model. Which approach best matches this requirement?
5. You deploy an Azure OpenAI model for a helpdesk assistant. You need to reduce the chance that the assistant generates harmful or inappropriate content. What should you implement?
This final chapter is where you convert “I’ve read the material” into “I can pass the exam.” AI-900 questions are rarely about deep math; they test whether you can recognize the right Azure AI workload, pick the correct service, and apply responsible AI and security constraints in realistic scenarios. Your job is to build a repeatable approach: pace yourself, identify what the question is really asking, and eliminate distractors based on service boundaries and core concepts.
You’ll work through two mock exam passes (Part 1 and Part 2), then perform a Weak Spot Analysis to target the areas that still produce hesitation. Finally, you’ll use an Exam Day Checklist to remove avoidable errors (misreads, timing panic, and configuration confusion). Throughout, map every decision back to the course outcomes: AI workloads, ML basics on Azure, computer vision, NLP, and generative AI with Copilot and Azure OpenAI—securely and responsibly.
Exam Tip: Treat every practice set as a “process drill.” A correct guess with a sloppy process doesn’t scale to exam pressure; a consistent process does.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 style questions often look straightforward, but they hide a “service boundary” or “workload type” decision. Before you start Mock Exam Part 1, set a pacing plan that prevents overthinking. Use a two-pass approach: Pass 1 answers everything you’re confident on immediately; Pass 2 revisits flagged items with a tighter elimination method.
In Pass 1, read the last line first (what are they asking you to choose?), then skim the scenario for constraints: data type (text, image, audio), requirement (classification vs extraction vs generation), operational need (real-time vs batch), and governance constraints (PII, region, private networking, content safety). AI-900 rewards accurate matching of requirement → service more than long reasoning.
Exam Tip: If a question contains “translate,” “transcribe,” “extract text,” “detect objects,” or “generate content,” circle the verb mentally; the verb often maps directly to a service family (Translator, Speech, OCR in Vision, Azure OpenAI).
Common pacing trap: spending too long on ML lifecycle details. The exam expects you to know the stages (data prep, training, evaluation, deployment, monitoring) and that Azure Machine Learning supports them, but it rarely expects implementation detail. If you find yourself debating hyperparameters, you’re probably off-objective. Flag it and move on.
Mock Exam Part 1 is designed to feel like real AI-900: scenario-heavy prompts that blend domains. You might see a business workflow described in plain language, and your job is to translate it into an Azure AI workload. Your first action is to classify the workload: ML prediction vs Cognitive Services-style prebuilt AI vs generative AI.
Scenario-heavy items often test “what is the workload” before “what is the service.” For example, a question may describe identifying products in photos (computer vision object detection), extracting fields from scanned forms (OCR + document understanding), detecting sentiment in support tickets (text analytics), or building a chat experience (conversational AI / question answering). When the scenario includes “create new text,” “summarize,” or “write code,” you are in generative AI territory, and you should consider Azure OpenAI patterns and guardrails.
Exam Tip: Watch for multi-step scenarios: “ingest documents, answer questions, and cite sources.” This is not just “chat”; it points to retrieval-augmented generation (RAG) thinking—grounding responses on enterprise data and controlling outputs through system instructions and content filtering.
Common traps in Part 1 include confusing classification with extraction. “Classify an image” means assign a label; “detect objects” means locate objects with bounding boxes; “OCR” means extract text. Another trap is assuming “machine learning” is always required; many problems are solved faster with prebuilt services (Vision, Speech, Language). When the scenario emphasizes custom model training and iterative improvement, then Azure Machine Learning becomes the stronger match.
As you complete Part 1, flag questions where you relied on intuition rather than a rule. Those become inputs to your Weak Spot Analysis later.
Mock Exam Part 2 shifts from narrative scenarios to service-selection pressure: multiple Azure options appear plausible, and only one best satisfies the exact requirement. Your strategy here is constraint matching. Identify the hard constraint (for example: “needs OCR,” “must translate,” “must generate,” “needs managed model hosting,” “requires responsible AI controls”), then eliminate services that cannot meet it by design.
Service-selection questions frequently test boundaries between Azure AI services: Vision vs Language vs Speech; Azure Machine Learning vs Azure OpenAI; and “prebuilt” vs “custom.” If the requirement is sentiment analysis, key phrase extraction, entity recognition, or language detection, the Language service is typically central. If it is speech-to-text, text-to-speech, or speaker recognition concepts, Speech is the anchor. For generative tasks (summarization, drafting, code generation), Azure OpenAI is the most direct fit, often paired with safety features and enterprise security controls.
Exam Tip: When you see “secure and responsible,” translate that into concrete controls: private endpoints/VNet integration where applicable, authentication/authorization via Azure AD, data governance, and content filtering (Azure AI Content Safety and Azure OpenAI safety features). The exam often rewards selecting the option that includes governance, not just capability.
Another frequent focus is ML lifecycle support: data labeling, experiment tracking, model registry, deployment endpoints, and monitoring. Azure Machine Learning maps cleanly to these. The trap is picking “a model API” when the prompt is about training and managing your own models. Conversely, the trap can be picking Azure Machine Learning when the prompt wants a prebuilt capability like OCR or translation.
After Part 2, you should have a clear list of “I mix these up” pairs (for example: OCR vs image classification; QnA vs chat; custom ML vs prebuilt AI; generative AI vs summarization in standard NLP). Those pairs will drive your review map.
Your Weak Spot Analysis begins with disciplined review. Don’t just note what you got wrong—diagnose why the wrong options were attractive. Use a four-step framework for every missed or guessed item: (1) restate the requirement in one sentence; (2) name the workload type (vision/NLP/speech/ML/generative); (3) identify the minimum service capability needed; (4) explain why each distractor fails a requirement or violates a constraint.
Distractors in AI-900 commonly fail in predictable ways. Some are “adjacent capability” distractors: a service in the right family but the wrong task (for instance, choosing image classification when bounding boxes are required). Others are “too generic” distractors: selecting “Azure Machine Learning” for a problem already solved by a prebuilt API. A third category is “governance blind” distractors: correct functionally, but missing security/responsible AI requirements.
Exam Tip: If two options both can technically solve the task, the exam usually wants the one that is (a) most direct, (b) least custom, and (c) aligned with Microsoft’s recommended service for that workload.
Also audit your reading errors. Many candidates lose points by ignoring a single word like “handwritten,” “real-time,” “custom,” “summarize,” or “private.” These words shift the service choice. Build a personal “trigger word” list from your mock exam review and rehearse it before the real exam.
This is your final review map: a compact set of memory anchors tied directly to the course outcomes and to what AI-900 tests. Start by anchoring each domain to a “verb → service” pattern. For AI workloads: identify whether the problem is prediction (ML), perception (vision/speech), language understanding (NLP), or generation (Azure OpenAI/Copilot patterns). For ML fundamentals: remember supervised vs unsupervised vs reinforcement learning at a high level, plus classification vs regression vs clustering, and that Azure Machine Learning supports end-to-end lifecycle (data, training, deployment, monitoring).
For computer vision, lock in three separations: classification (label), object detection (label + location), OCR (text extraction). Remember the exam can test “responsible vision” principles: privacy considerations, bias risks, and when to avoid or limit certain vision use cases. For NLP workloads: sentiment, entities, key phrases, translation, speech, and question answering all map to distinct service capabilities; don’t collapse them into one generic “AI service.”
Exam Tip: Generative AI questions often test safe usage patterns more than model trivia. Think: grounding data sources (RAG), prompt instructions (system vs user intent), content filtering, and protecting sensitive data. If the scenario is enterprise copilots, prefer answers that mention security boundaries and responsible AI controls.
Create a one-page “last-minute sheet” with these anchors and your personal confusion pairs from the mocks. Your goal is fast recognition under pressure, not relearning.
On exam day, eliminate preventable failure modes first: environment, timing, and mental state. Ensure your test setup is stable (device charged, reliable internet, quiet space if remote). If the exam is proctored, comply with workspace requirements early to avoid time loss. Plan to start with a calm buffer so you’re not carrying stress into the first questions.
Timing: commit to the two-pass method you practiced. In the first pass, answer and move. In the second pass, apply elimination using service boundaries and constraints. If you feel stuck, write (mentally or on provided scratch space) the key verb and data type; then ask “which Azure service is designed for this exact verb?”
Exam Tip: Anxiety often shows up as rereading the same paragraph repeatedly. When that happens, jump to the question prompt (what are they asking?), then scan for constraint words (custom, real-time, private, generate, detect, extract). This breaks the loop.
Final checklist items: verify you understand the difference between prebuilt AI services and custom ML, confirm your generative AI safety vocabulary (content safety, grounding, access control), and remind yourself that AI-900 is a breadth exam. You do not need deep implementation details—your score comes from correct workload recognition and correct service selection. Finish by reviewing flagged answers only if you have time; don’t churn on already-solid choices.
1. A company wants to build a chatbot that drafts responses to customer emails using the company’s product manuals. The manuals must not be used to train a public model, and access must be restricted using Azure identity. Which solution best meets the requirement?
2. You are reviewing practice exam results. You consistently miss questions asking you to choose between Azure OpenAI, Azure AI Language, and Azure AI Vision. Which Weak Spot Analysis action is most effective?
3. A team is building an app that must (1) identify objects in photos, (2) extract printed text from receipts, and (3) generate a short natural-language summary of the extracted text for the user. Which pairing of Azure services is the best fit?
4. During the exam, you notice a question with multiple requirements and tempting distractors. You have limited time remaining. Which approach best aligns with a repeatable AI-900 test-taking process?
5. A company wants to enable Microsoft Copilot features for employees. They are concerned about accidental exposure of sensitive internal data and want to reduce avoidable mistakes on rollout day. Which step is most appropriate to include in an Exam Day Checklist-style readiness checklist for this scenario?