HELP

+40 722 606 166

messenger@eduailast.com

AI-900 Generative AI: Copilot & Azure OpenAI Explained

AI Certification Exam Prep — Beginner

AI-900 Generative AI: Copilot & Azure OpenAI Explained

AI-900 Generative AI: Copilot & Azure OpenAI Explained

Master AI-900 domains with clear explanations and real exam-style practice.

Beginner ai-900 · microsoft · azure · azure-ai-fundamentals

Prepare for Microsoft AI-900 with a clear, domain-mapped blueprint

This Edu AI course is a focused exam-prep blueprint for the Microsoft AI-900: Azure AI Fundamentals exam—built for beginners with basic IT literacy and no prior certification experience. You’ll study exactly what the exam measures, in the same structure as the official domains, and you’ll practice with realistic, exam-style questions that train service selection, scenario reading, and elimination of distractors.

What this course covers (aligned to the official AI-900 domains)

The AI-900 exam expects you to recognize common AI solution patterns and map them to the right Azure services and concepts. This course is organized into six chapters that function like a short book: orientation, four domain-focused learning chapters, and a final mock exam with review.

  • Describe AI workloads: identify AI workload types (prediction, classification, detection, generation), when AI is appropriate, and core responsible AI ideas.
  • Fundamental principles of ML on Azure: understand supervised vs unsupervised learning, training vs inference, evaluation, and how Azure Machine Learning fits the lifecycle.
  • Computer vision workloads on Azure: choose services for image analysis, OCR, and common vision scenarios, with attention to privacy and responsible usage.
  • NLP workloads on Azure: select services for text analytics, translation, speech, question answering, and conversational solutions.
  • Generative AI workloads on Azure: explain LLM fundamentals, Copilot patterns, and how Azure OpenAI is used with grounding, safety, and governance considerations.

How the 6 chapters help you pass

Chapter 1 sets you up with the exam rules, registration path, scoring expectations, and a study strategy that works for busy schedules. Chapters 2–5 each focus on one or two official exam domains, combining conceptual understanding with frequent exam-style practice milestones. Chapter 6 provides a full mock exam split into two parts plus a structured review process to diagnose weak spots and tighten your timing.

Practice that mirrors the exam

AI-900 is not about coding—it’s about recognizing scenarios and choosing the best answer. The practice in this course emphasizes:

  • Reading for key requirements (latency, data type, constraints, outcomes)
  • Service matching (Vision vs Language vs Azure Machine Learning vs Azure OpenAI)
  • Responsible AI basics in real-world choices (privacy, fairness, transparency)
  • Common distractors and why they’re tempting but incorrect

Get started on Edu AI

If you’re new to Azure certifications, start by setting up your learning workflow and schedule. You can Register free to track progress and return to practice sets as you improve. Want to compare options? You can also browse all courses and pair this with other fundamentals paths.

Who this is for

This course is designed for individuals preparing for AI-900 who want a structured, domain-aligned plan. It’s ideal for students, career switchers, and IT professionals who need to understand Azure AI concepts—including Copilot and Azure OpenAI—at an exam-ready level.

What You Will Learn

  • Describe AI workloads and key considerations for choosing AI solutions on Azure (Describe AI workloads)
  • Explain core machine learning concepts and how Azure Machine Learning supports the ML lifecycle (Fundamental principles of ML on Azure)
  • Identify Azure services for image classification, object detection, OCR, and responsible vision solutions (Computer vision workloads on Azure)
  • Choose Azure services for text analytics, translation, speech, question answering, and conversational AI (NLP workloads on Azure)
  • Describe generative AI concepts, Copilot patterns, and how Azure OpenAI is used securely and responsibly (Generative AI workloads on Azure)

Requirements

  • Basic IT literacy (cloud concepts, web apps, and data basics)
  • No prior certification experience required
  • Ability to navigate a web browser and the Azure portal at a beginner level (optional but helpful)
  • Willingness to practice with exam-style multiple-choice and scenario questions

Chapter 1: AI-900 Exam Orientation and Study Plan

  • Understand what AI-900 measures and how domains are weighted
  • Register, schedule, and take the exam (online vs test center)
  • Scoring, question types, and time-management strategy
  • Build a 2-week and 4-week study plan with hands-on checkpoints

Chapter 2: Describe AI Workloads (and Responsible AI Basics)

  • Differentiate AI, ML, deep learning, and generative AI in scenarios
  • Match common workloads to Azure AI services at a high level
  • Apply responsible AI principles to real exam-style situations
  • Practice set: Describe AI workloads domain (exam-style questions)

Chapter 3: Fundamental Principles of Machine Learning on Azure

  • Understand supervised vs unsupervised learning and common algorithms
  • Explain training, validation, testing, and overfitting/underfitting
  • Walk through the Azure Machine Learning workspace and ML lifecycle
  • Practice set: ML principles on Azure (exam-style questions)

Chapter 4: Computer Vision Workloads on Azure

  • Identify vision tasks and pick the right Azure service per scenario
  • Understand OCR, image analysis, and document processing basics
  • Review security, privacy, and responsible vision considerations
  • Practice set: Computer vision workloads (exam-style questions)

Chapter 5: NLP and Generative AI on Azure (Copilot + Azure OpenAI)

  • Choose Azure services for sentiment, key phrases, NER, translation, and speech
  • Design conversational solutions: question answering and bots at a fundamentals level
  • Explain generative AI concepts and Copilot patterns for business scenarios
  • Use Azure OpenAI concepts: prompts, grounding, safety, and deployment basics
  • Practice set: NLP + Generative AI workloads (exam-style questions)

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist

Jordan Patel

Microsoft Certified Trainer (MCT)

Jordan Patel is a Microsoft Certified Trainer who helps beginners pass Microsoft Fundamentals exams with practical, scenario-first instruction. He has designed AI-900 exam-prep programs for teams and individuals, focusing on Azure AI services, responsible AI, and test-taking strategy.

Chapter 1: AI-900 Exam Orientation and Study Plan

This chapter sets your foundation for passing AI-900 with confidence and efficiency. AI-900 is a fundamentals exam, but it is not “just vocabulary.” Microsoft tests whether you can recognize real Azure AI workloads, pick the right service for the job, and apply responsible AI thinking—especially where generative AI and Copilot patterns appear in modern solutions.

Across this course you will build toward five outcomes: describe AI workloads and key considerations for choosing Azure AI solutions; explain core machine learning (ML) concepts and how Azure Machine Learning supports the ML lifecycle; identify Azure services for computer vision (classification, object detection, OCR, and responsible vision); choose Azure services for natural language processing (text analytics, translation, speech, QnA, and conversational AI); and describe generative AI concepts, Copilot patterns, and how Azure OpenAI is used securely and responsibly.

Use this chapter as your orientation and your plan. We’ll map the official objective areas to the way questions are written, clarify exam mechanics (registration, policies, question types, scoring), and then convert all of that into a 2-week and a 4-week schedule with hands-on checkpoints. Your goal is not to read more—it’s to retain more and recognize patterns under time pressure.

Practice note for Understand what AI-900 measures and how domains are weighted: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register, schedule, and take the exam (online vs test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Scoring, question types, and time-management strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a 2-week and 4-week study plan with hands-on checkpoints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand what AI-900 measures and how domains are weighted: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register, schedule, and take the exam (online vs test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Scoring, question types, and time-management strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a 2-week and 4-week study plan with hands-on checkpoints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand what AI-900 measures and how domains are weighted: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register, schedule, and take the exam (online vs test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: AI-900 overview and domain map (official objectives)

AI-900 (Microsoft Azure AI Fundamentals) measures your ability to describe AI workloads and identify Azure services that implement them. Expect heavy emphasis on “service selection” and “workload fit”: given a scenario, choose the Azure offering that solves it with the least complexity and the most alignment to requirements (latency, customization, data sensitivity, and responsible AI).

Microsoft periodically adjusts domain weightings, but the exam consistently covers: (1) AI workloads and considerations, (2) fundamental ML principles on Azure, (3) computer vision workloads, (4) NLP workloads, and (5) generative AI workloads—now including Copilot-style solution patterns and Azure OpenAI security/responsible use. When you study, organize notes by these domains, not by product marketing pages.

  • AI workloads: When to use rules vs ML vs generative models; supervised vs unsupervised vs reinforcement concepts; responsible AI basics.
  • ML on Azure: Training vs inference, features/labels, overfitting, evaluation metrics, and how Azure Machine Learning supports the lifecycle (data, training, deployment, monitoring).
  • Vision: Image classification vs object detection vs OCR; “responsible vision” and privacy-sensitive scenarios.
  • NLP: Text analytics, translation, speech-to-text/text-to-speech, question answering, and conversational AI patterns.
  • Generative AI: Prompts, grounding, retrieval-augmented generation (RAG) as a pattern, and using Azure OpenAI with security controls.

Exam Tip: Many questions are “compare and choose” where two options sound plausible. Anchor yourself by identifying the workload type first (vision/NLP/ML/genAI), then match to the simplest Azure service that meets requirements. Over-engineering (choosing Azure Machine Learning for a straightforward prebuilt API task) is a common trap.

Another frequent trap is confusing “train a model” with “use a pretrained model.” If the scenario needs custom training on your labeled data, that pushes you toward Azure Machine Learning or custom capabilities within Azure AI services. If the scenario is standard (OCR, language detection, sentiment), prebuilt Azure AI services are usually correct.

Section 1.2: Exam registration, policies, and ID requirements

You can take AI-900 through Pearson VUE either online (OnVUE proctoring) or at a test center. Registration is straightforward, but policy mistakes can cost you a forfeited attempt—treat logistics as part of your study plan.

Online delivery is convenient, but it is strict. You’ll be asked to photograph your testing area and present valid government-issued ID. Clear your desk completely, silence devices, and ensure a stable internet connection. Expect check-in steps and a waiting period. Test centers reduce the home-environment risk but require travel and scheduling flexibility.

Exam Tip: Choose the format that minimizes your personal risk. If you have unreliable internet, shared spaces, or frequent interruptions, schedule at a test center. A calm, predictable environment is worth more than the convenience of home testing.

  • ID and name matching: Ensure your Microsoft certification profile name matches your ID. Mismatches can prevent admission.
  • Rescheduling/cancellation: Know the reschedule window. Missing it can mean losing the fee.
  • Room rules (online): No second monitor, no visible notes, no phones, no smartwatches. Even reading questions aloud can be flagged.

Plan registration as a commitment device: pick a date that aligns with a 2-week or 4-week plan (see Section 1.5). The best candidates schedule early, then use the deadline to drive consistent practice rather than last-minute cramming.

Section 1.3: Exam format—MCQ, case studies, and scenario sets

AI-900 questions are designed to test recognition and decision-making under constraints. Expect multiple-choice and multiple-response items, plus scenario-based sets where several questions share the same context. Even when there is no formal “case study,” many items read like mini-stories: a business need, data type, and constraints, followed by “Which Azure service should you use?”

To perform well, read like an architect: highlight the nouns (image, audio, text, documents, chat) and the verbs (detect, classify, extract, translate, summarize, answer questions). Those words usually reveal the workload domain. Then look for qualifiers: “custom model,” “no-code,” “real-time,” “PII,” “on-premises data,” “responsible AI,” or “needs to cite sources.” Qualifiers are the difference between correct and almost-correct.

Exam Tip: Treat each question as a two-step problem: (1) identify the workload, (2) identify the service and the feature tier (prebuilt vs custom vs full ML). If you jump straight to product names, you’ll fall for distractors.

  • MCQ traps: Options that are real Azure products but not designed for the asked task (for example, using a general compute service when a specialized AI service is intended).
  • Multiple-response traps: One option is necessary but insufficient; another is sufficient alone. The exam often rewards “minimal complete solution.”
  • Scenario-set traps: Don’t forget earlier details. If the scenario mentions “must keep data within Azure” or “requires content filtering,” that should influence later answers about Azure OpenAI configuration and responsible AI.

Time management is part of format mastery. Don’t overinvest in early questions. Mark uncertain items for review, but avoid changing answers without a clear reason—most score drops come from second-guessing.

Section 1.4: Scoring model, pass expectations, and retake planning

Microsoft exams use a scaled scoring model. You don’t see a simple “X out of Y” score; you see a scaled score, and you need to meet the passing standard (commonly 700 on a 1000 scale, though the exact model is set by Microsoft). The key takeaway: not all questions necessarily contribute equally, and the difficulty distribution can vary by exam form.

Because of scaling, your strategy should be to maximize certainty on high-frequency objective areas. For AI-900, that typically means being able to confidently map scenarios to the correct Azure AI service and explain foundational ML concepts (training vs inference, evaluation, overfitting) without hesitation.

Exam Tip: Build your “must-not-miss list.” If you miss easy service-mapping items (OCR vs object detection; translation vs sentiment; Azure OpenAI vs Azure Machine Learning), you force yourself to be perfect on harder scenario questions—which is a losing trade.

Retake planning is also a professional skill. If you don’t pass, treat the score report as an objective map of weakness by domain. Do not restart from page one. Rebuild a 7–10 day remediation plan targeted to the lowest domain plus the second-lowest domain, and include hands-on reinforcement (labs) to prevent repeat confusion.

  • Pass expectations: You should be consistently scoring high on reputable practice sets and be able to justify every answer in terms of workload + requirement.
  • Retake mindset: Your second attempt should be faster and calmer; the goal is eliminating recurring traps, not learning everything again.

Finally, plan your exam date with a buffer. If your schedule is unpredictable, choose a date that still gives you at least a week of review flexibility without pushing you into cramming.

Section 1.5: Study strategy—active recall, spaced repetition, labs

AI-900 rewards candidates who can retrieve information quickly and apply it to scenarios. Passive reading feels productive but produces weak recall under exam pressure. Your core strategy should combine active recall (forcing your brain to answer without looking), spaced repetition (revisiting at increasing intervals), and hands-on labs (turning abstract services into memorable workflows).

Exam Tip: If you can’t explain why an option is wrong, you don’t fully own the concept. Train with “justify the distractor”: practice naming the scenario where each wrong option would be correct.

Use two timelines depending on your availability:

  • 2-week plan (accelerated): Days 1–3: AI workloads + responsible AI + service map overview. Days 4–6: ML fundamentals + Azure Machine Learning lifecycle. Days 7–9: Vision + NLP (one day each, one day mixed practice). Days 10–11: Generative AI + Azure OpenAI security/responsibility + Copilot patterns. Days 12–14: full mixed review, error log focus, and one timed practice session.
  • 4-week plan (steady): Week 1: AI workloads + responsible AI, daily flashcards. Week 2: ML fundamentals + Azure ML, two short labs. Week 3: Vision and NLP, alternating days with scenario practice. Week 4: Generative AI + Azure OpenAI + comprehensive review and timed practice.

Hands-on checkpoints keep your studying honest. At minimum, schedule: (1) one Azure AI Vision or OCR trial, (2) one Azure AI Language task (sentiment/translation or entity recognition), (3) one Azure Machine Learning “hello world” (even a conceptual walk-through of training/deployment), and (4) one Azure OpenAI studio experience focused on prompt design and safety controls. The goal is not depth; it is building mental anchors so exam choices feel obvious.

Section 1.6: Using Microsoft Learn + Edu AI practice workflow

Microsoft Learn provides the official-aligned learning paths, terminology, and service boundaries. Use it as your source of truth for what Microsoft intends you to know—especially around updated Azure AI service names, responsible AI commitments, and generative AI guidance. Your workflow should connect Learn content to repeated retrieval practice and targeted correction.

Use a simple loop:

  • Learn: Read one objective-focused unit (e.g., OCR vs object detection; training vs inference; Azure OpenAI responsible use).
  • Recall: Close the page and write a 5–7 line summary from memory: what it is, when to use it, and what it is not.
  • Apply: Convert your summary into scenario cues: “If I see X requirement, I choose Y.”
  • Verify: Do a short practice block and check explanations carefully. Add misses to an error log categorized by domain.

Exam Tip: Keep an “Azure service boundary table” as a living document. Many AI-900 errors come from boundary confusion—mixing Azure Machine Learning (custom model lifecycle) with Azure AI services (prebuilt and some customization) and with Azure OpenAI (generative models, prompt-based behavior, safety tooling).

Integrate Edu AI practice as your feedback engine: after each practice session, tag each miss with (1) domain, (2) concept type (definition vs scenario mapping vs responsible AI), and (3) the specific keyword you missed (e.g., “OCR,” “custom,” “PII,” “grounding”). Then schedule a spaced repetition review: same day quick fix, 48-hour re-test, and end-of-week mixed set. This workflow turns practice into measurable progress rather than repeated guessing.

By the end of this chapter, your objective is simple: you should know what the exam measures, how you will take it, how you will manage time and scoring expectations, and exactly how your 2-week or 4-week plan will produce retention through labs and targeted practice.

Chapter milestones
  • Understand what AI-900 measures and how domains are weighted
  • Register, schedule, and take the exam (online vs test center)
  • Scoring, question types, and time-management strategy
  • Build a 2-week and 4-week study plan with hands-on checkpoints
Chapter quiz

1. You are planning your AI-900 preparation and want to prioritize study time based on how Microsoft measures the skills in the exam. Which approach aligns best with the AI-900 exam orientation and objective-based preparation strategy?

Show answer
Correct answer: Use the official AI-900 skills outline (objective domains) as a checklist and allocate study time proportionally to domain weighting, validating learning with hands-on checkpoints.
AI-900 is a fundamentals exam, but it tests more than vocabulary: you must recognize AI workloads, select appropriate Azure AI services, and apply responsible AI concepts. Using the official objective domains and weighting helps you invest time where the exam focuses, and hands-on checkpoints improve pattern recognition. Memorizing definitions alone is insufficient because service selection and workload recognition are assessed. Focusing only on generative AI/Copilot is also incorrect because the exam spans multiple domains (AI workloads, ML concepts, vision, NLP, and generative AI), not a single area.

2. A candidate is deciding whether to take AI-900 online or at a test center. They are concerned about avoiding exam-day disruptions and policy violations. Which action best reduces the risk of an invalidated exam attempt?

Show answer
Correct answer: Schedule either delivery option, but ensure you can meet the delivery-specific policies (for example, a compliant testing environment for online proctoring or valid ID and arrival time for a test center).
The orientation for exam mechanics emphasizes following delivery-specific policies. Online proctoring typically prohibits additional monitors, phones, and reference materials; violating these rules can invalidate the exam. Test centers also enforce strict rules (ID checks, security procedures, timing), so assuming they are less strict is incorrect. The correct approach is choosing the delivery mode that fits you while adhering to its policies.

3. You are taking AI-900 and notice that some items are scenario-based and others are short knowledge checks. You want a time-management strategy that aligns with the exam’s question types and scoring approach. What should you do?

Show answer
Correct answer: Answer straightforward questions quickly, flag time-consuming scenario questions for review, and maintain a steady pace because the exam measures broad objective coverage under time pressure.
AI-900 includes a mix of question types and evaluates your ability to recognize patterns across objective areas within a time limit. A realistic strategy is to move efficiently through easy items, flag lengthy ones, and return if time permits. Over-investing time early risks incomplete sections. Skipping scenario questions is incorrect because scenario-based items commonly contribute to the score and frequently test service selection and responsible AI considerations.

4. A company wants to build an internal Copilot-like assistant that drafts emails and summarizes meetings using company data. During AI-900 preparation, which learning outcome should you prioritize to be most exam-ready for this scenario?

Show answer
Correct answer: Describe generative AI concepts, Copilot patterns, and how Azure OpenAI is used securely and responsibly.
The scenario maps directly to the AI-900 outcome focused on generative AI concepts, Copilot patterns, and responsible/secure use of Azure OpenAI. AI-900 is not a developer exam requiring building neural networks from scratch or implementing OCR algorithms; instead, it emphasizes recognizing workloads and selecting appropriate Azure AI services and patterns. OCR may be relevant in other scenarios, but it is not the primary match for a Copilot-style summarization/drafting assistant.

5. You have two weeks to prepare for AI-900 and want a plan that matches the chapter’s recommended approach. Which study plan is most aligned with the course guidance?

Show answer
Correct answer: Follow a 2-week schedule that covers each objective domain, includes periodic hands-on checkpoints (labs/demos), and ends with timed practice questions to build pattern recognition under exam constraints.
The chapter emphasizes efficiency: retain more by mapping objectives to how questions are written, using hands-on checkpoints, and practicing under time pressure. Reading without practice reduces your ability to recognize real Azure AI workloads. Skipping responsible AI is incorrect because AI-900 evaluates key considerations and responsible AI thinking in Azure AI solutions, including modern generative AI patterns.

Chapter 2: Describe AI Workloads (and Responsible AI Basics)

This chapter targets the AI-900 “Describe AI workloads” objective and the exam’s recurring expectation: you can look at a scenario and correctly name the workload (prediction vs classification vs detection vs generation), then select the most appropriate Azure service family at a high level. You are not expected to design deep architectures, but you are expected to recognize patterns (vision, language, speech, decisioning) and apply responsible AI basics to common deployment situations.

As you study, practice “keyword-to-workload” mapping: words like forecast, estimate, probability often point to prediction/regression; which category suggests classification; find where suggests detection; create new text/images/code points to generative AI. The exam frequently tests whether you can avoid overcomplicating: many questions are solved by picking the right workload type and the right service family, not by selecting a specific model algorithm.

Exam Tip: If a prompt describes “extracting text from images,” do not label it as generative AI. That is OCR (a computer vision extraction workload). Generative AI creates new content; OCR converts existing content from one representation to another.

Practice note for Differentiate AI, ML, deep learning, and generative AI in scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match common workloads to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to real exam-style situations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: Describe AI workloads domain (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate AI, ML, deep learning, and generative AI in scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match common workloads to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to real exam-style situations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: Describe AI workloads domain (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate AI, ML, deep learning, and generative AI in scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match common workloads to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: AI workloads—prediction, classification, detection, generation

Section 2.1: AI workloads—prediction, classification, detection, generation

AI-900 expects you to distinguish common AI workload types based on the business question being asked. Most scenarios fit into one of four buckets: prediction, classification, detection, or generation. Your goal is to identify what the system must output.

Prediction (often regression) outputs a numeric value or probability. Typical phrasing: “predict sales next month,” “estimate time to failure,” “forecast call volume,” or “predict likelihood of churn.” The output is not a label like “fraud/not fraud,” but a number (or a probability that can later be thresholded).

Classification outputs a discrete label or category. Examples: “is this email spam?”, “which product category?”, “positive/neutral/negative sentiment,” “safe/unsafe content.” Multi-class classification chooses one of many categories; binary classification chooses between two. On the exam, classification is frequently confused with detection—watch for “where” vs “what.”

Detection identifies the presence and often the location of an item or event. In vision, object detection returns bounding boxes (where the object is in the image). In security/IoT, anomaly detection identifies unusual patterns in time-series data. Many candidates incorrectly call anomaly detection “classification,” but the scenario usually emphasizes “unusual/outlier” rather than a fixed label set.

Generation creates new content: text, code, images, summaries, or embeddings used for semantic search. Phrasing: “draft an email,” “summarize a document,” “create a job description,” “generate product images,” “write code from a prompt.” The key is novelty—content that didn’t exist before.

Exam Tip: If the system returns both a label and an explanation, it is still classification; explanation does not change the workload type. If it returns a label plus a bounding box, that’s detection.

Section 2.2: When to use AI vs rules-based logic (decision criteria)

Section 2.2: When to use AI vs rules-based logic (decision criteria)

Another exam skill is deciding whether AI is appropriate at all. Not every “automation” scenario needs machine learning or generative AI. Rules-based logic (if/then, lookup tables, deterministic workflows) is often faster, cheaper, and easier to govern—especially when requirements are stable and you can precisely define the logic.

Use rules-based logic when: the domain is deterministic; inputs are structured and limited; the organization needs full explainability; or the cost of errors is extremely high and rules can cover all cases. Example: “Reject expense claims over $5,000 without manager approval.” No model is required.

Use AI/ML when: the rules are hard to write; patterns are subtle; inputs are unstructured (text, images, audio); or the environment changes. Example: “Detect fraudulent transactions” or “classify customer emails by intent.” ML can learn patterns that are impractical to hand-code.

Use generative AI when: the goal is to produce natural language or creative output, transform text (summarize, rewrite), or enable conversational experiences that need flexible responses. But generative AI introduces extra governance needs (prompt injection, hallucinations, data leakage), so the exam may push you to pair it with responsible controls.

Common trap: Choosing generative AI for simple retrieval. If a question only needs “return the company policy paragraph,” a search or QnA approach can be more appropriate than free-form generation. A generative model may still be used, but the scenario must justify it (summarization, reasoning, conversational UX, or grounding with retrieved content).

Exam Tip: If the requirement says “must always be correct and follow strict rules,” lean rules-based or constrained workflows. If it says “unstructured input” or “too many variations,” lean AI.

Section 2.3: Core Azure AI service families (overview mapping)

Section 2.3: Core Azure AI service families (overview mapping)

At a high level, the exam focuses on recognizing Azure service families rather than memorizing every SKU. When you see vision, language, speech, or generative needs, map them to the correct family, then select the best fit within that family if the question asks.

Azure AI Services (formerly Cognitive Services) are prebuilt APIs for common workloads. This includes Vision, Speech, and Language capabilities (translation, sentiment, key phrases, named entity recognition, etc.). If a scenario describes “call an API to analyze an image or text” with minimal training, Azure AI Services is a strong default.

Azure Machine Learning supports the full ML lifecycle: data prep, training, tracking, deployment, and monitoring. Choose it when the scenario emphasizes building a custom model, training on the organization’s data, or MLOps (pipelines, model registry, endpoints). This is a major differentiator: prebuilt AI Services vs custom ML in Azure ML.

Azure OpenAI Service is used for generative AI with large language models and image generation models (depending on availability). Scenarios: chat assistants, summarization, content generation, code generation, embeddings for semantic search. The exam often expects you to mention secure usage: authentication, network controls, and responsible filters.

Azure AI Search (often paired with Azure OpenAI) supports indexing and retrieval over enterprise content. If the scenario mentions “search across documents,” “semantic ranking,” or “ground responses in internal PDFs,” think AI Search as the retrieval component—even if generation is done by Azure OpenAI.

Common trap: Picking Azure Machine Learning for simple OCR or translation. If no custom training is needed and a prebuilt API exists, the exam usually expects Azure AI Services. Conversely, if the scenario explicitly says “train a model with your labeled images,” Azure Machine Learning (and/or custom vision capabilities) becomes more relevant.

Exam Tip: Look for verbs: “train,” “experiment,” “deploy model endpoint,” “monitor drift” → Azure Machine Learning. “Analyze image/text,” “extract,” “translate,” “transcribe” → Azure AI Services. “Generate/summarize/chat/embeddings” → Azure OpenAI.

Section 2.4: Responsible AI concepts (fairness, reliability, privacy)

Section 2.4: Responsible AI concepts (fairness, reliability, privacy)

AI-900 includes responsible AI at a conceptual level: you must recognize risks and the appropriate principle to apply. Three commonly tested principles are fairness, reliability & safety, and privacy & security. The exam does not expect legal expertise; it expects correct identification of the concern and a sensible mitigation direction.

Fairness means the system should not produce biased outcomes for groups defined by sensitive attributes (for example, gender, age, ethnicity) or proxies. Exam scenarios often describe different error rates across groups (higher false rejects for one group). The correct response is to evaluate bias, improve representative data, and monitor model performance by subgroup.

Reliability and safety means the system behaves consistently under expected conditions and fails safely. In generative AI, this includes reducing harmful outputs and ensuring grounded responses. In traditional ML, it includes robustness to data changes and careful threshold selection (e.g., false positives vs false negatives in medical screening).

Privacy and security covers protecting personal and confidential data: least privilege access, encryption, data minimization, and preventing unintended disclosure. For generative AI, privacy concerns include prompts containing sensitive data and outputs leaking confidential information. For vision, privacy concerns include facial recognition use and storing images longer than necessary.

Common trap: Confusing privacy with transparency. If the issue is “users don’t know how decisions are made,” that’s transparency. If the issue is “data could expose personal information,” that’s privacy.

Exam Tip: When a scenario mentions protected classes or unequal treatment, answer with fairness. When it mentions outages, unsafe behavior, or inconsistent results, answer with reliability/safety. When it mentions PII, consent, retention, or access controls, answer with privacy/security.

Section 2.5: Transparency, inclusiveness, accountability in Azure contexts

Section 2.5: Transparency, inclusiveness, accountability in Azure contexts

The remaining responsible AI principles frequently appear in scenario form: transparency, inclusiveness, and accountability. The exam wants you to connect these principles to practical actions in Azure-based solutions.

Transparency means stakeholders understand what the system can and cannot do. Practically: communicate that outputs may be probabilistic; provide explanations where possible; document data sources and limitations; and label AI-generated content when appropriate. In Azure, transparency is often addressed through clear user experience design (disclosures), model documentation, and logging for traceability.

Inclusiveness means the solution is accessible and usable by people with diverse abilities and backgrounds. Scenario cues: speech systems struggling with accents, captions needed for hearing-impaired users, vision models missing certain lighting/skin tones, or a chatbot that fails non-native language phrasing. The mitigation is broader testing, representative datasets, and fallback user experiences (human handoff, alternative input methods).

Accountability means humans remain responsible for outcomes. You typically implement governance: role assignments, review processes, audit logs, and escalation paths. For generative AI copilots, accountability often means “human in the loop” approval before actions are taken (sending emails, updating records) and monitoring for misuse.

Common trap: Treating accountability as “the model is accurate.” Accuracy is performance; accountability is ownership and oversight. If the scenario mentions “who is responsible” or “approval/audit,” it’s accountability.

Exam Tip: If the prompt says “users rely on outputs as facts,” transparency is the missing principle—add disclosures, citations/grounding, and guidance on verification. If it says “the system must work for all users,” inclusiveness. If it says “ensure oversight and auditability,” accountability.

Section 2.6: Practice questions—Describe AI workloads (domain-focused)

Section 2.6: Practice questions—Describe AI workloads (domain-focused)

This domain is often scored through short scenarios where two answers seem plausible. Your strategy is to (1) identify the workload type, (2) select the Azure family that best matches the level of customization, and (3) apply responsible AI if the scenario includes risk cues.

First, underline the output: number/probability (prediction), label (classification), location + label (detection), extracted text (OCR extraction), or new content (generation). Then check whether training is required. If the scenario says “use a prebuilt API,” “no data science team,” or “minimal setup,” choose Azure AI Services. If it says “train on our labeled dataset,” “track experiments,” or “deploy our own model,” choose Azure Machine Learning. If it says “draft, summarize, chat, or generate,” choose Azure OpenAI (often with retrieval patterns using Azure AI Search when internal documents are involved).

Next, scan for responsible AI triggers. Mentions of unequal impact across groups should immediately shift you to fairness considerations. Mentions of sensitive data, customer records, or regulations should shift you to privacy/security controls (access, encryption, data minimization). Mentions of unsafe behavior, unpredictable outputs, or critical decisions should trigger reliability/safety plus human oversight.

Exam Tip: Many wrong answers are “near misses” (right service, wrong workload). For example, selecting a generative service when the task is sentiment analysis (classification) is a common distractor. If the scenario is about “analyze,” “extract,” or “detect,” default to non-generative workloads unless it explicitly asks to create new content.

Finally, remember the exam’s level: it tests recognition, not implementation detail. You rarely need to name a specific model architecture. If you can confidently say “this is classification using a prebuilt language API” or “this is generation using Azure OpenAI with governance controls,” you are answering at the expected AI-900 depth.

Chapter milestones
  • Differentiate AI, ML, deep learning, and generative AI in scenarios
  • Match common workloads to Azure AI services at a high level
  • Apply responsible AI principles to real exam-style situations
  • Practice set: Describe AI workloads domain (exam-style questions)
Chapter quiz

1. A retail company wants to estimate next month’s sales for each store using historical sales, promotions, and holidays. Which AI workload is being described?

Show answer
Correct answer: Regression (prediction)
Estimating a numeric value (next month’s sales) is a prediction/regression workload. Classification would assign labels (for example, "high/medium/low demand") rather than producing a numeric forecast. Object detection is a computer vision workload focused on finding and locating items in images or video, which is not part of this scenario.

2. You need to extract printed text from scanned invoices and store it as searchable text. Which Azure AI service family is the best fit at a high level?

Show answer
Correct answer: Azure AI Vision (OCR)
Extracting text from images is OCR, a computer vision extraction workload, typically handled by Azure AI Vision capabilities. Azure OpenAI is used to generate new text or transform text with generative models; it does not primarily perform OCR on images. Azure AI Speech is for speech-to-text and text-to-speech, not for reading text from scanned documents.

3. A manufacturer wants to identify whether each product coming off an assembly line is "defective" or "not defective" based on sensor readings. What workload is this?

Show answer
Correct answer: Classification
Choosing between discrete categories ("defective" vs "not defective") is a classification workload. Regression predicts a continuous numeric output (for example, defect probability as a number can be output, but the business goal described is a categorical decision). Text generation is a generative AI workload for creating new natural language content, which is unrelated to sensor-based defect labeling.

4. A city wants an AI solution that locates pedestrians and bicycles in traffic camera footage by drawing bounding boxes around them. Which workload is being described?

Show answer
Correct answer: Object detection
Drawing bounding boxes and finding the location of objects is object detection. Image classification would label an entire image (for example, "contains a bicycle") without returning coordinates. Sentiment analysis is a language workload that evaluates opinions/emotions in text, not objects in video.

5. A bank deploys an AI model to help approve loans. After release, it discovers approval rates differ significantly across demographic groups. Which Responsible AI principle is most directly being violated?

Show answer
Correct answer: Fairness
Disparate outcomes for different demographic groups most directly indicates a fairness issue (bias and unequal treatment/impact). Privacy and security concerns focus on protecting data and preventing unauthorized access, which is not what the scenario highlights. Reliability and safety focuses on consistent, dependable performance and avoiding harmful failures; while important, the core issue described is inequitable outcomes across groups.

Chapter 3: Fundamental Principles of Machine Learning on Azure

This chapter maps directly to the AI-900 objective area Fundamental principles of ML on Azure: you must explain core machine learning concepts (features, labels, training vs inference, overfitting/underfitting) and connect them to how Azure Machine Learning (Azure ML) supports the ML lifecycle. The exam is not asking you to derive math; it is asking you to recognize ML workload types, choose the right approach, and identify Azure ML components that enable training and deployment.

Expect scenario-style questions: “You have historical labeled data…,” “You don’t have labels…,” “Your model performs well in training but poorly in production…,” or “You need repeatable training and deployment….” Your job is to translate those phrases into the correct learning type and the correct Azure ML capability.

Exam Tip: On AI-900, the fastest path to correct answers is vocabulary matching. If you see label, think supervised. If you see group similar, think clustering. If you see deploy for predictions, think endpoint/inference. If you see repeatable workflow, think pipeline.

The sections below walk through supervised vs unsupervised learning, training/validation/testing and generalization, and a practical tour of Azure ML concepts (workspace, compute, pipelines, endpoints). The chapter ends with a practice set section (no questions embedded here) to guide what you should be able to do under exam time pressure.

Practice note for Understand supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain training, validation, testing, and overfitting/underfitting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Walk through the Azure Machine Learning workspace and ML lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: ML principles on Azure (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain training, validation, testing, and overfitting/underfitting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Walk through the Azure Machine Learning workspace and ML lifecycle: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: ML principles on Azure (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand supervised vs unsupervised learning and common algorithms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: ML basics—features, labels, inference, and evaluation metrics

Section 3.1: ML basics—features, labels, inference, and evaluation metrics

Machine learning problems on the AI-900 exam are described using a small set of core terms. Features are the input variables (columns) used to make a prediction: age, temperature, transaction amount, image pixels, or text embeddings. A label (also called target) is what you want to predict: churn yes/no, house price, defect category, or fraud vs not fraud. When you run a trained model on new data to produce outputs, that is inference.

The exam frequently checks whether you can pick the right evaluation metric for the objective. For classification, you will see accuracy, precision, recall, and F1 score. Accuracy is overall correctness but can be misleading for imbalanced classes (for example, 99% non-fraud). Precision answers “when the model predicts positive, how often is it right?” Recall answers “of all actual positives, how many did we catch?” For regression, look for MAE (mean absolute error), MSE/RMSE, and R-squared—metrics that measure how close numeric predictions are to actual values.

  • Classification metrics: accuracy, precision, recall, F1; use confusion matrix language (TP/FP/TN/FN) to reason about tradeoffs.
  • Regression metrics: MAE/RMSE for error magnitude; R-squared for explained variance.

Exam Tip: If the scenario emphasizes “minimize false alarms” (false positives), lean toward improving precision. If it emphasizes “don’t miss any” (false negatives), lean toward improving recall. These words often appear verbatim in item stems.

Common trap: mixing up the “output” of a model with the “label.” Labels are known during training; outputs are produced during inference. Another trap is assuming “accuracy” is always the best metric. If the question mentions rare events, skewed data, or compliance/safety risks, accuracy alone is often not the best choice.

Section 3.2: Supervised learning scenarios (regression and classification)

Section 3.2: Supervised learning scenarios (regression and classification)

Supervised learning means you have labeled examples and you want the model to learn a mapping from features to labels. AI-900 expects you to identify whether the label is numeric (regression) or categorical (classification), then match the scenario to a typical algorithm family. You are not tested on implementation details, but you are tested on recognizing what fits.

Regression predicts a number: forecasting demand, estimating delivery time, predicting energy usage, or valuing a property. Common algorithms include linear regression and decision tree–based regression. Classification predicts a category: spam vs not spam, defect type A/B/C, customer segment label, or disease positive/negative. Common algorithms include logistic regression, decision trees, and support vector machines; in modern practice you may also see boosted trees and neural networks, but the exam tends to stay conceptual.

  • Regression keywords: predict, estimate, forecast, continuous value, numeric output.
  • Classification keywords: category, class, yes/no, label, discrete outcome.

Exam Tip: If the prompt includes “probability of …” and the outcome is yes/no, it is still classification (often logistic regression conceptually). Probability outputs do not automatically make it regression.

Another frequent exam pattern is mapping the business goal to the correct learning type. “Predict whether a customer will churn” is classification; “predict how many days until churn” is regression. Also watch the trap where a scenario includes text or images: the data type can be unstructured, but the learning task is still classification/regression depending on the label.

Finally, supervised learning depends on data quality: label accuracy, feature relevance, and representativeness. If the question hints that labels are missing or unreliable, supervised learning may be inappropriate—or you may need more data preparation and validation controls before training.

Section 3.3: Unsupervised learning scenarios (clustering and anomaly detection)

Section 3.3: Unsupervised learning scenarios (clustering and anomaly detection)

Unsupervised learning is used when you do not have labels and you want to discover structure in the data. On AI-900, the two most testable unsupervised patterns are clustering and anomaly detection. Clustering groups similar items together—customer segmentation, grouping documents by topic, or organizing product catalogs. A classic algorithm is k-means, but you mainly need to recognize the “group similar without labels” intent.

Anomaly detection looks for unusual behavior compared to a baseline: detecting suspicious transactions, equipment sensor spikes, network intrusions, or rare manufacturing defects. The exam will often describe anomalies as “outliers,” “unexpected patterns,” or “deviations from normal.” Importantly, anomaly detection can be unsupervised (no labeled fraud/normal) or semi-supervised (trained mostly on normal data). Your clue is whether labeled examples of the abnormal class exist.

  • Clustering keywords: segment, group, similarity, discover categories, no labels.
  • Anomaly keywords: outlier, rare event, unusual, deviation, alerting.

Exam Tip: If the scenario asks you to “create segments” or “group by similarity,” do not choose classification just because the output looks like a category. Classification requires known labeled categories; clustering creates groups based on the data.

Common trap: confusing anomaly detection with binary classification. If the question explicitly says you have labeled examples of fraud vs not fraud, classification is viable. If it says “we don’t know what fraud looks like yet” or “no historical labels,” anomaly detection is the better conceptual fit.

Unsupervised methods are also used for exploratory analysis before supervised training, such as discovering new customer segments that later become labels. The exam may test this indirectly by asking which approach helps “understand” data rather than “predict” an outcome.

Section 3.4: Model training lifecycle—data prep, training, validation, testing

Section 3.4: Model training lifecycle—data prep, training, validation, testing

The ML lifecycle is a frequent AI-900 theme: prepare data, train a model, evaluate it, and deploy it for inference. The exam checks whether you can distinguish training, validation, and testing and recognize overfitting and underfitting symptoms.

Data preparation includes cleaning (handling missing values), transforming (encoding categories), normalizing/scaling when appropriate, and splitting data into sets. Training fits model parameters using the training set. Validation is used to tune hyperparameters and select a model configuration. Testing is a final, unbiased evaluation on unseen data to estimate real-world performance.

Overfitting occurs when a model learns training noise: high training performance, low validation/test performance. Underfitting occurs when the model is too simple or not trained enough: poor performance on both training and validation/test. Questions often describe these patterns in plain language (for example, “works great on training data but fails in production”).

  • Overfitting signs: training accuracy high, test accuracy low; model too complex; insufficient regularization; data leakage.
  • Underfitting signs: both accuracies low; model too simple; features insufficient; training not converged.

Exam Tip: Watch for data leakage, a common trap: if the same customer appears in both training and test, or if a feature indirectly contains the label (like “closed_date” when predicting “will close”), the model will look unrealistically good. The exam may not use the term “leakage,” but it will describe suspiciously perfect metrics.

Another trap is mixing up validation and testing. Validation is for tuning during development; testing is the final check. If a question asks which set you use to “choose hyperparameters,” that is validation. If it asks which set provides the “final estimate” of performance, that is test.

Section 3.5: Azure Machine Learning concepts—compute, pipelines, endpoints

Section 3.5: Azure Machine Learning concepts—compute, pipelines, endpoints

Azure Machine Learning (Azure ML) is Azure’s primary service for managing the ML lifecycle at scale. For AI-900, you should be able to describe what an Azure ML workspace is and identify key building blocks: compute, data, experiments, pipelines, and endpoints. The exam typically stays at the “what is it for?” level, not SDK syntax.

A workspace is the top-level container that organizes assets (datasets, models, jobs, endpoints) and ties into Azure security and governance. Compute is where work runs. You will see compute instances (interactive dev) and compute clusters (scalable training). The practical exam cue: if the question needs scale-out training, choose a cluster; if it needs a personal dev machine in the cloud, choose an instance.

Pipelines represent repeatable workflows: data prep steps, training steps, evaluation, and registration. They matter for consistency and MLOps-style automation. Endpoints are for deployment/inference. A real-time endpoint supports low-latency online predictions; batch inference is for scoring large datasets asynchronously. The exam may simply say “deploy the model so apps can call it,” which points to endpoints.

  • Compute instance: interactive development, notebooks, single VM experience.
  • Compute cluster: scalable training, distributed jobs, cost control via autoscaling.
  • Pipeline: repeatable multi-step workflow (prep → train → evaluate).
  • Endpoint: managed deployment surface for inference (real-time or batch).

Exam Tip: Separate “training” words from “serving” words. Train, tune, experiment → compute/pipelines. Deploy, consume, predict, call from an app → endpoint/inference.

Common trap: selecting Azure AI services (prebuilt) when the prompt clearly needs custom model training. If the scenario says you have your own labeled dataset and want to train and deploy a custom model, Azure ML is the expected direction.

Section 3.6: Practice questions—Fundamental principles of ML on Azure

Section 3.6: Practice questions—Fundamental principles of ML on Azure

This section is your checklist for exam readiness on “Fundamental principles of ML on Azure.” Even when questions are wrapped in business context, they usually reduce to a few decisions: supervised vs unsupervised, classification vs regression, which dataset split is being discussed, what metric matters, and which Azure ML component enables the requirement.

When you do an exam-style practice set, force yourself to underline the clue words in the stem: “labeled,” “predict a number,” “group similar,” “outlier,” “hyperparameter tuning,” “deploy,” “low latency,” “batch scoring,” “repeatable workflow.” Then map each clue to a concept: learning type, metric, or Azure ML asset.

  • Be able to justify why a scenario is classification vs regression using the label type, not the data type.
  • Be able to diagnose overfitting vs underfitting from training vs validation/test behavior.
  • Be able to choose validation for tuning and test for final evaluation.
  • Be able to match compute instance, compute cluster, pipeline, and endpoint to the scenario requirement.

Exam Tip: If two answers both sound “ML-ish,” pick the one that matches the operational requirement. Example patterns: “repeatable and auditable training” points to pipelines; “scale training” points to clusters; “app calls for predictions” points to endpoints.

Common traps to watch during practice: (1) choosing accuracy for imbalanced problems, (2) treating clustering results as “labels” and calling it classification, (3) using the test set to tune, and (4) confusing training infrastructure (compute) with deployment infrastructure (endpoint). If you can consistently avoid these, you are aligned with what AI-900 typically tests in this chapter’s objective area.

Chapter milestones
  • Understand supervised vs unsupervised learning and common algorithms
  • Explain training, validation, testing, and overfitting/underfitting
  • Walk through the Azure Machine Learning workspace and ML lifecycle
  • Practice set: ML principles on Azure (exam-style questions)
Chapter quiz

1. A retail company has two years of transaction history where each record includes customer attributes (age, region, average basket value) and a known outcome indicating whether the customer churned. The company wants to predict churn for current customers. Which machine learning approach should you use?

Show answer
Correct answer: Supervised learning (classification)
This is supervised learning because the dataset contains a label (churned: yes/no) and the goal is to predict that label for new customers, which is a classification task. Clustering is unsupervised and would only group customers by similarity without learning a churn label. Anomaly detection is typically used to find unusual patterns (e.g., fraud) rather than predict a known categorical outcome.

2. You train a model in Azure Machine Learning and it achieves very high accuracy on the training dataset but performs significantly worse on the test dataset. What is the most likely issue?

Show answer
Correct answer: Overfitting
High training performance with poor test performance indicates overfitting: the model learned training-specific noise rather than general patterns. Underfitting would usually show poor performance on both training and test data due to an overly simple model. High training accuracy alone does not demonstrate generalization; generalization is validated by performance on unseen validation/test data.

3. A data science team wants to tune hyperparameters for a model in Azure Machine Learning. They need a dataset split that they can use repeatedly during model selection without using the final unbiased dataset reserved for evaluation. Which dataset split should they primarily use for this purpose?

Show answer
Correct answer: Validation set
The validation set is used during training iterations for model selection and hyperparameter tuning. The test set should be held back for final, unbiased evaluation after tuning is complete; using it during tuning risks leaking information and inflating metrics. Training-only evaluation can hide overfitting and does not provide an independent check while tuning.

4. A company is building an ML solution on Azure and wants a repeatable process that prepares data, trains the model, and registers the resulting model in a consistent way each time the workflow runs. Which Azure Machine Learning capability best supports this requirement?

Show answer
Correct answer: Azure Machine Learning pipelines
Azure Machine Learning pipelines are designed to orchestrate and automate repeatable ML workflows (data preparation, training, evaluation, and model registration). Endpoints are used for deployment/inference (serving predictions) rather than orchestrating training steps. Dashboards provide visibility/monitoring but do not define a repeatable end-to-end training workflow.

5. A manufacturer has sensor readings from machines but no labels indicating failure types. They want to group machines into sets with similar operating behavior to identify patterns. Which technique should they use?

Show answer
Correct answer: Clustering (unsupervised learning)
Because there are no labels and the goal is to group similar records, clustering is the appropriate unsupervised approach. Classification requires labeled categories to learn from (e.g., specific failure types). Regression also requires labeled numeric targets (e.g., remaining useful life) and is not intended for grouping by similarity.

Chapter 4: Computer Vision Workloads on Azure

This chapter maps directly to the AI-900 objective area Computer vision workloads on Azure: identifying vision tasks, choosing the right Azure service for image classification, object detection, OCR, and responsible vision solutions. On the exam, you are rarely asked to write code—you are tested on whether you can recognize the workload (what problem is being solved) and select the correct Azure capability or service family for that workload.

Expect scenario-style questions: “A retail company wants to…” or “A healthcare provider needs to…” Your job is to translate the scenario into a vision task (classification vs detection vs OCR vs document understanding), then choose the best-fit Azure service. The most common trap is picking a tool based on a keyword (“image,” “document,” “text”) rather than the actual output required (labels, bounding boxes, extracted text, or structured fields).

We’ll walk through core vision task types, Azure AI Vision capabilities, and how OCR differs from document processing. We’ll also cover security/privacy and responsible vision patterns—an increasingly tested area because it affects deployment decisions, not just model accuracy.

Practice note for Identify vision tasks and pick the right Azure service per scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand OCR, image analysis, and document processing basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Review security, privacy, and responsible vision considerations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: Computer vision workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify vision tasks and pick the right Azure service per scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand OCR, image analysis, and document processing basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Review security, privacy, and responsible vision considerations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: Computer vision workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify vision tasks and pick the right Azure service per scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand OCR, image analysis, and document processing basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Vision workload types—classification, detection, segmentation

AI-900 expects you to distinguish the “shape” of the output for common computer vision tasks. This is the fastest way to identify the correct answer in a scenario question.

Image classification assigns one (or more) labels to an entire image. The output is a label and a confidence score, such as “dog: 0.98” or “damaged-package: 0.87.” Use classification when the question is about deciding what the image is overall.

Object detection identifies and localizes objects in an image. The output includes labels plus bounding boxes (coordinates). Use detection when the question asks where items are, how many there are, or if you must draw boxes around items (people, cars, defects on a surface).

Segmentation is more granular than detection: it assigns a class to each pixel (or produces a mask) to outline objects precisely. Segmentation is used when bounding boxes are not precise enough—for example, measuring the area of a tumor region, calculating the amount of spilled liquid, or separating foreground/background.

  • How the exam tests this: A scenario will hint at the required output (“identify the breed” → classification; “count all products on shelf” → detection; “highlight the damaged region” → segmentation).
  • Common trap: Choosing classification when the question implies multiple objects in a single image. If you need counts or locations, you need detection/segmentation.
  • Exam Tip: Look for verbs: “classify,” “categorize,” “tag” → classification; “detect,” “locate,” “count,” “draw boxes” → detection; “mask,” “outline,” “pixel-level” → segmentation.

In this course’s “pick the right Azure service per scenario” lesson, you’ll repeatedly apply this translation step: first identify the vision task, then map to the Azure capability that produces the required output.

Section 4.2: Azure AI Vision capabilities (image analysis and OCR concepts)

Azure vision questions typically revolve around Azure AI Vision capabilities that provide prebuilt analysis of images and videos. The exam focus is not on the API surface, but on what you can extract: tags, captions, objects, and text.

Image analysis refers to general-purpose understanding of image content. You may see features like generating tags (keywords), creating a caption/description, detecting common objects, or analyzing image properties. These are best when you don’t want to train a custom model and the goal is broad metadata enrichment (for search, accessibility, or content organization).

OCR (optical character recognition) extracts printed or handwritten text from images. On AI-900, OCR is frequently a distinct requirement: “extract text from a photo,” “read serial numbers,” “digitize scanned pages,” or “capture text from a storefront sign.” OCR output is text plus layout-related hints (lines/words), depending on the capability.

  • How to identify OCR in scenarios: The scenario involves images but the desired output is text (not labels like “receipt” or “invoice”).
  • Common trap: Confusing OCR with document understanding. OCR extracts text; document processing extracts structured fields (like total, vendor, date) and may use OCR as one component.

Exam Tip: If the question asks for “extract text” or “read text,” OCR is the primary feature. If it asks for “extract invoice number” or “total amount,” that’s a document processing workload (Section 4.3), even though OCR will likely be involved behind the scenes.

This section ties to the lesson “Understand OCR, image analysis, and document processing basics.” For the exam, practice separating “free-form text extraction” (OCR) from “field extraction into named properties” (document intelligence).

Section 4.3: Document processing overview (forms, receipts, IDs as scenarios)

Document processing is a specialized subset of vision workloads where the goal is not merely to read text, but to convert documents into structured data. AI-900 typically frames this as forms processing, invoice/receipt extraction, and ID document scenarios.

In a forms scenario, the user might submit scanned forms with checkboxes, typed entries, or handwritten fields. The expected output is a set of key-value pairs (for example: “EmployeeName=…”, “StartDate=…”) and often table extraction. In receipts, you’ll commonly see fields like merchant name, date, subtotal, tax, and total. In IDs (where allowed by policy and region), the scenario may ask to extract fields such as name, date of birth, or document number.

  • What the exam is really asking: “Do you need raw text, or do you need fields?” Raw text → OCR. Fields/tables → document processing.
  • Common trap: Selecting a general image analysis capability because the input is an image. The question is judged by the output requirement: structured extraction is the key phrase.
  • Exam Tip: Words like “invoice,” “receipt,” “form,” “fields,” “table,” “key-value pairs,” “structured” are strong indicators of document processing rather than general OCR.

From an architecture perspective, document processing often includes validation steps (for example, cross-check totals, verify required fields) and can feed downstream systems like CRM/ERP. On AI-900, you won’t be tested on integration mechanics, but you may be tested on selecting the correct service category for “automate data entry from documents.”

Section 4.4: Custom vs prebuilt vision solutions (when customization matters)

A major decision point in Azure vision workloads is whether to use prebuilt capabilities or build a custom model. The exam tests your ability to choose customization when the scenario contains domain-specific requirements that prebuilt models can’t reliably satisfy.

Prebuilt solutions are ideal when the problem matches common patterns: general tagging, captions, common object detection, or generic OCR/document extraction. They are faster to implement, require little to no training data, and are typically cost-effective for proof-of-concepts.

Custom vision solutions matter when you must recognize company-specific items (custom product SKUs, proprietary components, specialized medical imagery, unusual defect types) or when you need to enforce your own label taxonomy. Customization usually implies you will supply labeled images and iterate to achieve required accuracy.

  • Signals for custom solutions in exam questions: “unique to our business,” “proprietary parts,” “not covered by existing labels,” “high accuracy required on specialized images,” or “must detect our specific defect categories.”
  • Common trap: Choosing custom when the scenario doesn’t demand it. If the task is generic (e.g., “detect people wearing hard hats”), a prebuilt option may be sufficient depending on the wording. The exam often rewards the simplest solution that meets requirements.
  • Exam Tip: If the scenario mentions “limited labeled data” and a tight timeline, that usually pushes you toward prebuilt. If it mentions “new classes,” “domain-specific,” or “must be trained,” that pushes you toward custom.

In “Identify vision tasks and pick the right Azure service per scenario,” your fastest method is: (1) identify task type, (2) decide prebuilt vs custom based on domain specificity and labeling needs, (3) select the service family aligned to that output.

Section 4.5: Responsible vision—PII, bias risk, and human-in-the-loop patterns

AI-900 increasingly emphasizes responsible AI considerations. For vision workloads, this commonly shows up as questions about privacy, PII, consent, and human review for sensitive decisions. You should assume that images can contain personal data (faces, license plates, ID cards, medical information) even if the scenario does not explicitly call it out.

PII and security considerations: Minimize data collection, store images securely, apply encryption, and restrict access via least privilege. In many architectures, you should avoid storing raw images unless necessary; store derived results (extracted fields, labels) when that meets the requirement. Redaction or blurring is a common mitigation when images are used for analytics but contain sensitive regions.

Bias and fairness risk: Vision models can underperform across lighting conditions, skin tones, camera types, or geographic contexts. The exam won’t require a deep statistical treatment, but it will test whether you recognize that you must evaluate model performance across representative samples and monitor drift over time.

Human-in-the-loop (HITL) patterns: For high-impact outcomes (identity verification, safety incidents, compliance flags), build a review workflow where uncertain predictions or low-confidence results are routed to a human. Combine confidence thresholds with audit logs to support accountability.

  • Common trap: Treating vision output as ground truth. On the exam, “fully automated decision-making” is often the wrong choice when the scenario involves sensitive personal data or safety/compliance impacts.
  • Exam Tip: When you see “ID documents,” “customers,” “patients,” “employees,” or “surveillance,” think: consent, retention policy, access control, and HITL for exceptions.

This section aligns to the lesson “Review security, privacy, and responsible vision considerations.” If an answer choice mentions governance actions (auditability, access controls, human review, minimizing data), it is often the best-fit in responsible AI questions.

Section 4.6: Practice questions—Computer vision workloads on Azure

This chapter’s practice set (provided separately) is designed to mirror AI-900’s scenario style without requiring implementation details. To score well, focus on the decision framework the exam expects: identify the workload, match the output type, then pick the simplest Azure service that satisfies requirements.

Step-by-step method for exam questions:

  • Step 1: Identify the input modality. Is the input an image/photo, a scanned document, or a camera stream? If it’s a document, ask whether the goal is text or fields.
  • Step 2: Identify the output shape. Label (classification), box (detection), mask (segmentation), text (OCR), or structured fields/tables (document processing).
  • Step 3: Decide prebuilt vs custom. Generic labels → prebuilt; proprietary categories → custom.
  • Step 4: Apply responsible constraints. If PII/sensitive use cases are involved, look for answers that include minimization, access control, and/or human review.

Exam Tip: When two answers both “sound plausible,” choose the one that matches the required output most precisely. AI-900 favors correct workload mapping over broader, catch-all descriptions.

Common traps to watch for in practice: (1) Selecting OCR when the scenario asks for named fields like totals and dates; (2) selecting general image tagging when the scenario needs localization/counting; (3) selecting custom training when the scenario explicitly wants a quick deployment using prebuilt capabilities; (4) ignoring privacy/HITL requirements in scenarios involving IDs or individuals.

Use the practice set to drill pattern recognition: under timed conditions, you should be able to classify the scenario into one of the few core categories in under 15–20 seconds, then spend the remaining time eliminating distractors.

Chapter milestones
  • Identify vision tasks and pick the right Azure service per scenario
  • Understand OCR, image analysis, and document processing basics
  • Review security, privacy, and responsible vision considerations
  • Practice set: Computer vision workloads (exam-style questions)
Chapter quiz

1. A retail company wants to analyze in-store camera images to identify whether a shelf is empty and highlight the empty shelf area in the image. Which computer vision task and Azure service best fit this requirement?

Show answer
Correct answer: Object detection using Azure AI Vision
The requirement is to both identify the condition and locate the region (highlight an area), which aligns to object detection (bounding boxes/regions). Image classification returns labels for the whole image but does not localize objects/areas. OCR is for extracting text and is not appropriate for detecting empty shelf regions.

2. A healthcare provider scans patient intake forms that contain printed text and handwritten entries. They need the text extracted from the scans, but they do not need to map values to specific fields. Which Azure capability should you use?

Show answer
Correct answer: OCR (Read) with Azure AI Vision
This scenario is text extraction from images/documents without requiring structured field mapping, which is an OCR/Read workload. Object detection is used to locate objects (e.g., syringes, devices) with bounding boxes, not to extract text content. Image classification provides an overall label (e.g., 'form' vs 'not form') but does not extract the text.

3. An insurance company receives thousands of claim documents. They must extract structured information such as policy number, claimant name, and claim amount from specific parts of each document. Which Azure service is the best fit?

Show answer
Correct answer: Azure AI Document Intelligence
Extracting structured fields from documents (key-value pairs/tables) is a document processing workload, best handled by Azure AI Document Intelligence. Azure AI Vision image analysis and OCR can extract text and describe images, but they are not optimized for reliably mapping extracted text into structured fields like policy number or claim amount. Object detection is for locating visual objects, not document field extraction.

4. A company wants to classify product images as 'acceptable' or 'damaged' before listing them online. The images show products on a plain background, and no bounding boxes are required. Which approach should you choose?

Show answer
Correct answer: Image classification using Azure AI Vision
The goal is to assign a label to the entire image ('acceptable' vs 'damaged'), which is image classification. Object detection is unnecessary because the requirement does not include locating specific damaged regions with bounding boxes. OCR is unrelated because no text extraction is required.

5. A financial services company plans to process customer-submitted ID photos to extract text and verify details. They want to reduce privacy risk and support responsible AI practices. Which design choice best aligns with security and privacy considerations for vision workloads on Azure?

Show answer
Correct answer: Minimize data retention by processing images and storing only the extracted results when possible
A key responsible AI and privacy practice is data minimization: retain only what you need (for example, store extracted text/fields rather than raw images) and limit retention duration. Storing images indefinitely increases privacy and compliance risk. Disabling access controls contradicts security best practices; Azure services support securing resources via authentication/authorization and should not be weakened for convenience.

Chapter 5: NLP and Generative AI on Azure (Copilot + Azure OpenAI)

This chapter maps directly to AI-900 objectives around NLP workloads on Azure and Generative AI workloads on Azure. The exam expects you to recognize common text, translation, speech, and conversational scenarios and choose the correct Azure service (often by spotting keywords like “sentiment,” “entity extraction,” “translation,” “speech-to-text,” “chatbot,” or “grounding with enterprise data”). You are not expected to design a production architecture, but you are expected to pick the right managed capability and describe what it does at a high level.

We’ll start with how NLP workloads are categorized (classification, extraction, summarization, and translation), then connect those workloads to Azure AI Language and Speech services. Finally, we’ll shift to generative AI: what LLMs do, why tokens/embeddings matter, and how Azure OpenAI + Copilot patterns use grounding/RAG and safety controls. Throughout, focus on how the exam phrases requirements and how to eliminate distractors that mention the wrong service family or an overly complex option.

Exam Tip: On AI-900, “choose the service” questions are usually won by matching the task verb (analyze, extract, translate, transcribe, answer, generate) to the correct Azure AI capability and ignoring unrelated platform components (Kubernetes, networking, data lakes) unless the prompt explicitly asks about them.

Practice note for Choose Azure services for sentiment, key phrases, NER, translation, and speech: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Design conversational solutions: question answering and bots at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts and Copilot patterns for business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Use Azure OpenAI concepts: prompts, grounding, safety, and deployment basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: NLP + Generative AI workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose Azure services for sentiment, key phrases, NER, translation, and speech: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Design conversational solutions: question answering and bots at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts and Copilot patterns for business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Use Azure OpenAI concepts: prompts, grounding, safety, and deployment basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: NLP workloads—classification, extraction, summarization, translation

NLP workloads on the AI-900 exam are typically framed by what you want to do with text: classification (assign labels), extraction (pull structured info out), summarization (shorten content while preserving meaning), and translation (convert languages). The exam often gives a business story (support tickets, reviews, emails, call transcripts) and asks which Azure service or feature to use.

Classification includes sentiment analysis (positive/negative/neutral), topic labeling, or categorizing documents by type. ExtractionSummarizationTranslation

On Azure, these tasks generally map to Azure AI services rather than building and training a model from scratch. For AI-900, treat managed APIs as the default answer unless the question explicitly says you must train a custom model or you have highly specialized labels that require custom training.

  • Text analytics-style tasks: sentiment, key phrases, entity recognition, language detection.
  • Translation: Azure AI Translator.
  • Summarization: can be delivered through language capabilities (extractive/abstractive summarization) or, increasingly in exam scenarios, via generative AI with Azure OpenAI when the prompt says “generate” or “compose.”

Common trap: Confusing “summarize” as a classic NLP feature vs. generative summarization. If the question emphasizes “natural-sounding, rewritten summary” or “draft executive brief,” that points to generative AI (Azure OpenAI). If it emphasizes “extract key sentences” or “extract highlights,” it can point to language summarization features. Use the wording: “generate” and “draft” are strong generative cues.

Exam Tip: When you see “extract entities,” “key phrases,” or “sentiment,” think Azure AI Language first. When you see “translate,” think Translator. When you see “transcribe,” think Speech. When you see “write,” “compose,” or “create,” think Azure OpenAI.

Section 5.2: Azure AI Language fundamentals (sentiment, NER, key phrases)

Azure AI Language is the exam’s core service family for text analysis. At a fundamentals level, you should be able to describe what the common features do and when you would choose them: sentiment analysis, named entity recognition (NER), and key phrase extraction. These features help transform unstructured text into structured signals for dashboards, routing, and automation.

Sentiment analysis evaluates the emotional tone of text (for example, customer reviews). In exam scenarios, look for language like “determine whether feedback is positive or negative” or “measure customer satisfaction from surveys.” NER identifies real-world entities (people, organizations, locations, dates, product names) and is often used for indexing, compliance, and search. Key phrase extraction pulls the most important terms from a document; it’s commonly used for tagging support tickets or summarizing themes at scale.

The exam also expects you to understand that these are prebuilt capabilities—fast to adopt, minimal ML expertise required. If a question says the organization needs a custom taxonomy (for example, domain-specific categories), then “custom text classification” can be a better fit than generic sentiment or key phrases.

Common trap: Confusing NER with key phrase extraction. NER returns entities with types (Person, Location, Organization), while key phrases returns important terms without entity typing. If the question requires “identify names and addresses” or “find company names,” that’s NER (and potentially PII detection). If it requires “identify main topics” or “auto-tag keywords,” that’s key phrases.

How to identify the correct answer: underline the noun being extracted. If it’s “entities” (names, places), choose NER. If it’s “keywords,” choose key phrases. If it’s “tone/opinion,” choose sentiment.

Exam Tip: If a scenario mentions compliance, privacy, or redaction (for example, “remove credit card numbers before storing text”), think PII detection within language capabilities rather than sentiment/NER alone—AI-900 often tests your ability to spot the governance requirement hidden in the story.

Section 5.3: Speech and conversational AI concepts (speech-to-text, bots)

Speech and conversational AI questions test whether you can distinguish between speech processing (turn audio into text or text into audio) and conversation orchestration (managing dialogues, intents, and responses). On Azure, speech-to-text and text-to-speech are classic Speech service capabilities, while bots and Q&A experiences use conversational building blocks.

Speech-to-text appears when the prompt mentions “transcribe calls,” “convert meeting audio to text,” or “create captions.” Text-to-speech

For conversational solutions, the exam frequently contrasts two approaches: (1) question answering over a knowledge base (FAQ-style) and (2) bots that manage multi-turn interactions. A Q&A solution is appropriate when users ask factual questions and expect consistent answers sourced from curated content. A bot becomes relevant when the conversation requires state, branching logic, or integration actions (e.g., “reset my password,” “check my order status”).

Common trap: Treating speech-to-text as the same as “language understanding.” Speech-to-text only produces text; you may still need language analysis or a conversational layer to interpret intent or answer questions. Another trap is assuming every chatbot must be generative—many exam questions still expect a deterministic Q&A system for predictable support answers.

Exam Tip: If the scenario says “users ask questions based on internal documents/FAQs,” pick a question answering capability. If it says “users need a conversational interface that can perform tasks,” pick a bot framework approach. If it says “audio,” always consider Speech first before language or OpenAI.

Section 5.4: Generative AI basics—LLMs, tokens, embeddings, prompt patterns

Generative AI questions on AI-900 validate that you understand what an LLM is and what it is not. An LLM (large language model) predicts the next token in a sequence, enabling it to generate text, summarize, rewrite, classify, and answer in natural language. The exam focuses on foundational terms: tokens, embeddings, and common prompt patterns.

Tokens are chunks of text used by the model (not necessarily words). Token limits affect how much input plus output you can fit into a single request, which matters when the scenario mentions “long documents” or “conversation history.” Embeddings are vector representations of meaning; they enable semantic search and similarity matching. When the prompt says “find the most relevant passages” or “search by meaning, not keywords,” embeddings are the concept being tested.

Prompt patterns show up as “how do you instruct the model?” Common patterns include: providing clear instructions, giving examples (few-shot), setting a role, specifying output format, and adding constraints (tone, length, citations). For exam purposes, you should recognize that better prompts reduce ambiguity and help produce consistent outputs, but prompts are not a security boundary.

Common trap: Overstating determinism. LLM outputs can vary; even with strong prompts, generative results are probabilistic. If the scenario requires strict, repeatable outputs or guaranteed factuality, the correct design often includes grounding with trusted data and/or deterministic components rather than “just prompt the model.”

Exam Tip: Watch for phrasing like “semantic similarity,” “vector,” “closest match,” or “retrieve relevant content”—that is your clue to embeddings and retrieval patterns, not classic keyword search and not sentiment/NER.

Section 5.5: Azure OpenAI + Copilot concepts—grounding, RAG, safety filters

Azure OpenAI is Microsoft’s managed offering for using OpenAI models with Azure enterprise controls. The exam wants you to understand secure and responsible usage patterns: deployment basics, prompting, grounding, RAG (retrieval augmented generation), and safety filters. You typically won’t be asked for API syntax; you will be asked to choose the right approach and describe what it accomplishes.

Deployments in Azure OpenAI represent configured model endpoints in your Azure subscription. Exam questions may use the word “deploy” to mean “make a model available for inference with specific settings.” This is different from deploying an app service or VM.

Grounding means anchoring responses in trusted, relevant information—often your organization’s data—so the model is less likely to hallucinate and more likely to answer with correct context. The most common grounding pattern is RAG: retrieve relevant passages (often using embeddings + vector search) and provide them to the model as context for generation. In Copilot-style business scenarios (HR policy assistant, IT helpdesk, contract Q&A), RAG is the typical pattern because it keeps the model up to date without retraining.

Safety and responsible AI show up in requirements like “prevent harmful content,” “filter hate/violence/sexual content,” “reduce jailbreak risk,” or “protect sensitive data.” Azure OpenAI includes content filtering and monitoring options, but the exam expects you to know the high-level idea: you apply safety controls and you design to minimize exposure of sensitive information. Also recognize that grounding improves factuality but does not automatically guarantee compliance; you still need access controls and data governance.

Common trap: Thinking RAG “trains the model on your data.” RAG does not fine-tune by default; it retrieves and injects context at runtime. If a question explicitly asks to “adapt the model to a domain with new behavior and consistent tone,” that leans toward fine-tuning (though AI-900 usually emphasizes the concept more than the mechanics).

Exam Tip: When you see “must answer using our documents,” “must cite sources,” or “reduce hallucinations,” pick grounding/RAG. When you see “block unsafe content,” pick safety filters/content moderation controls. When you see “enterprise secure access,” emphasize Azure-managed deployments and governance rather than public, unmanaged endpoints.

Section 5.6: Practice questions—NLP workloads + Generative AI workloads on Azure

This chapter’s practice set should train your “service selection reflex,” which is exactly what AI-900 tests. The exam questions are rarely about deep implementation; they are about matching a requirement to the correct workload and Azure service family while avoiding distractors that sound plausible but don’t meet the requirement.

When you review practice items, use a consistent elimination method:

  • Identify the input modality: text vs. audio. Audio strongly suggests Speech (transcription or synthesis) before anything else.
  • Identify the verb: analyze/extract/translate/transcribe/answer/generate. The verb is usually the fastest path to the correct service.
  • Check for “enterprise constraints”: safety, PII, compliance, “use our data,” or “must be grounded.” Those constraints often push the answer toward Azure OpenAI with RAG and safety controls rather than generic NLP.
  • Decide if the output must be deterministic: FAQ answers and policy responses often favor question answering over free-form generation unless the question explicitly wants drafting or creative language.

Also watch how questions try to bait you with overlapping terminology. “Summarize” could mean extractive highlights (language features) or generative rewriting (Azure OpenAI). “Chatbot” could mean a bot with orchestrated dialogs, or a Q&A assistant, or a generative assistant with grounding. Your job is to match the business need: consistent factual answers (Q&A), task completion (bot), or drafting and flexible language (generative).

Exam Tip: If the scenario mentions “internal knowledge base,” “company policies,” “product manuals,” or “SharePoint content,” assume grounding is expected. If it also mentions “reduce hallucinations” or “cite sources,” that is a strong RAG signal. If it mentions “detect sentiment,” “extract entities,” or “key phrases,” that is a strong Azure AI Language signal.

Finally, remember that AI-900 often tests responsible AI posture implicitly. If two answers both seem to work, choose the one that includes safety, governance, and using managed services appropriately—those clues are frequently the tie-breaker.

Chapter milestones
  • Choose Azure services for sentiment, key phrases, NER, translation, and speech
  • Design conversational solutions: question answering and bots at a fundamentals level
  • Explain generative AI concepts and Copilot patterns for business scenarios
  • Use Azure OpenAI concepts: prompts, grounding, safety, and deployment basics
  • Practice set: NLP + Generative AI workloads (exam-style questions)
Chapter quiz

1. A retail company wants to analyze customer reviews to determine whether each review is positive, negative, neutral, or mixed. Which Azure service should you use?

Show answer
Correct answer: Azure AI Language (Sentiment analysis)
Sentiment analysis is an Azure AI Language capability designed to classify text by sentiment. Azure AI Translator focuses on language translation, not sentiment scoring. Azure AI Speech converts audio to text (or text to audio) and does not perform sentiment classification on text.

2. You need to extract people, organizations, and locations from support tickets stored as text. Which Azure capability best fits this requirement?

Show answer
Correct answer: Azure AI Language (Named Entity Recognition)
Named Entity Recognition (NER) in Azure AI Language is used to identify and extract entities such as person, organization, and location. Text to speech generates audio from text and does not extract entities. Document translation translates content between languages but does not identify entities within the text.

3. A travel app must translate short user messages from French to English in real time. Which Azure service should you choose?

Show answer
Correct answer: Azure AI Translator
Azure AI Translator is the managed service for text translation between languages. Key phrase extraction identifies main points in text but does not translate. Azure OpenAI can generate text and may be prompted to translate, but the exam typically expects the dedicated translation service for straightforward translation requirements.

4. A company wants to build a customer-facing chatbot that answers questions using content from their internal policy documents. They want responses grounded in that data rather than purely generated from the model. Which approach best matches this requirement?

Show answer
Correct answer: Use Azure OpenAI with grounding/RAG over the policy documents
Grounding (often implemented with RAG) with Azure OpenAI is used to answer questions by retrieving relevant enterprise content and using it to produce responses aligned to the source documents. Speech transcription does not provide question answering grounded in documents. Translation changes language but does not create a conversational Q&A experience tied to retrieved internal content.

5. You deploy an Azure OpenAI model for a helpdesk assistant. You need to reduce the chance that the assistant generates harmful or inappropriate content. What should you implement?

Show answer
Correct answer: Azure OpenAI content filtering/safety controls
Azure OpenAI includes safety features such as content filtering designed to help detect and mitigate harmful content. Language detection only identifies the language of input text and does not provide safety controls. Speaker recognition is a speech scenario feature and is unrelated to moderating generated text.

Chapter 6: Full Mock Exam and Final Review

This final chapter is where you convert “I’ve read the material” into “I can pass the exam.” AI-900 questions are rarely about deep math; they test whether you can recognize the right Azure AI workload, pick the correct service, and apply responsible AI and security constraints in realistic scenarios. Your job is to build a repeatable approach: pace yourself, identify what the question is really asking, and eliminate distractors based on service boundaries and core concepts.

You’ll work through two mock exam passes (Part 1 and Part 2), then perform a Weak Spot Analysis to target the areas that still produce hesitation. Finally, you’ll use an Exam Day Checklist to remove avoidable errors (misreads, timing panic, and configuration confusion). Throughout, map every decision back to the course outcomes: AI workloads, ML basics on Azure, computer vision, NLP, and generative AI with Copilot and Azure OpenAI—securely and responsibly.

Exam Tip: Treat every practice set as a “process drill.” A correct guess with a sloppy process doesn’t scale to exam pressure; a consistent process does.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam instructions and pacing strategy (AI-900 style)

AI-900 style questions often look straightforward, but they hide a “service boundary” or “workload type” decision. Before you start Mock Exam Part 1, set a pacing plan that prevents overthinking. Use a two-pass approach: Pass 1 answers everything you’re confident on immediately; Pass 2 revisits flagged items with a tighter elimination method.

In Pass 1, read the last line first (what are they asking you to choose?), then skim the scenario for constraints: data type (text, image, audio), requirement (classification vs extraction vs generation), operational need (real-time vs batch), and governance constraints (PII, region, private networking, content safety). AI-900 rewards accurate matching of requirement → service more than long reasoning.

Exam Tip: If a question contains “translate,” “transcribe,” “extract text,” “detect objects,” or “generate content,” circle the verb mentally; the verb often maps directly to a service family (Translator, Speech, OCR in Vision, Azure OpenAI).

Common pacing trap: spending too long on ML lifecycle details. The exam expects you to know the stages (data prep, training, evaluation, deployment, monitoring) and that Azure Machine Learning supports them, but it rarely expects implementation detail. If you find yourself debating hyperparameters, you’re probably off-objective. Flag it and move on.

  • Timebox each flagged question to a short revisit window.
  • Assume the exam is testing “best fit,” not “anything that could work.”
  • Don’t ignore responsible AI constraints; they can be the deciding factor.
Section 6.2: Mock Exam Part 1 (mixed domains, scenario-heavy)

Mock Exam Part 1 is designed to feel like real AI-900: scenario-heavy prompts that blend domains. You might see a business workflow described in plain language, and your job is to translate it into an Azure AI workload. Your first action is to classify the workload: ML prediction vs Cognitive Services-style prebuilt AI vs generative AI.

Scenario-heavy items often test “what is the workload” before “what is the service.” For example, a question may describe identifying products in photos (computer vision object detection), extracting fields from scanned forms (OCR + document understanding), detecting sentiment in support tickets (text analytics), or building a chat experience (conversational AI / question answering). When the scenario includes “create new text,” “summarize,” or “write code,” you are in generative AI territory, and you should consider Azure OpenAI patterns and guardrails.

Exam Tip: Watch for multi-step scenarios: “ingest documents, answer questions, and cite sources.” This is not just “chat”; it points to retrieval-augmented generation (RAG) thinking—grounding responses on enterprise data and controlling outputs through system instructions and content filtering.

Common traps in Part 1 include confusing classification with extraction. “Classify an image” means assign a label; “detect objects” means locate objects with bounding boxes; “OCR” means extract text. Another trap is assuming “machine learning” is always required; many problems are solved faster with prebuilt services (Vision, Speech, Language). When the scenario emphasizes custom model training and iterative improvement, then Azure Machine Learning becomes the stronger match.

As you complete Part 1, flag questions where you relied on intuition rather than a rule. Those become inputs to your Weak Spot Analysis later.

Section 6.3: Mock Exam Part 2 (mixed domains, service-selection focus)

Mock Exam Part 2 shifts from narrative scenarios to service-selection pressure: multiple Azure options appear plausible, and only one best satisfies the exact requirement. Your strategy here is constraint matching. Identify the hard constraint (for example: “needs OCR,” “must translate,” “must generate,” “needs managed model hosting,” “requires responsible AI controls”), then eliminate services that cannot meet it by design.

Service-selection questions frequently test boundaries between Azure AI services: Vision vs Language vs Speech; Azure Machine Learning vs Azure OpenAI; and “prebuilt” vs “custom.” If the requirement is sentiment analysis, key phrase extraction, entity recognition, or language detection, the Language service is typically central. If it is speech-to-text, text-to-speech, or speaker recognition concepts, Speech is the anchor. For generative tasks (summarization, drafting, code generation), Azure OpenAI is the most direct fit, often paired with safety features and enterprise security controls.

Exam Tip: When you see “secure and responsible,” translate that into concrete controls: private endpoints/VNet integration where applicable, authentication/authorization via Azure AD, data governance, and content filtering (Azure AI Content Safety and Azure OpenAI safety features). The exam often rewards selecting the option that includes governance, not just capability.

Another frequent focus is ML lifecycle support: data labeling, experiment tracking, model registry, deployment endpoints, and monitoring. Azure Machine Learning maps cleanly to these. The trap is picking “a model API” when the prompt is about training and managing your own models. Conversely, the trap can be picking Azure Machine Learning when the prompt wants a prebuilt capability like OCR or translation.

After Part 2, you should have a clear list of “I mix these up” pairs (for example: OCR vs image classification; QnA vs chat; custom ML vs prebuilt AI; generative AI vs summarization in standard NLP). Those pairs will drive your review map.

Section 6.4: Answer review framework—why the distractors are wrong

Your Weak Spot Analysis begins with disciplined review. Don’t just note what you got wrong—diagnose why the wrong options were attractive. Use a four-step framework for every missed or guessed item: (1) restate the requirement in one sentence; (2) name the workload type (vision/NLP/speech/ML/generative); (3) identify the minimum service capability needed; (4) explain why each distractor fails a requirement or violates a constraint.

Distractors in AI-900 commonly fail in predictable ways. Some are “adjacent capability” distractors: a service in the right family but the wrong task (for instance, choosing image classification when bounding boxes are required). Others are “too generic” distractors: selecting “Azure Machine Learning” for a problem already solved by a prebuilt API. A third category is “governance blind” distractors: correct functionally, but missing security/responsible AI requirements.

Exam Tip: If two options both can technically solve the task, the exam usually wants the one that is (a) most direct, (b) least custom, and (c) aligned with Microsoft’s recommended service for that workload.

Also audit your reading errors. Many candidates lose points by ignoring a single word like “handwritten,” “real-time,” “custom,” “summarize,” or “private.” These words shift the service choice. Build a personal “trigger word” list from your mock exam review and rehearse it before the real exam.

  • Was the mistake a workload misclassification (vision vs NLP vs generative)?
  • Was it a capability mismatch (classification vs detection vs OCR)?
  • Was it a lifecycle mismatch (training/manage vs consume API)?
  • Was it a governance miss (PII, safety, access control)?
Section 6.5: Final domain review map (last-minute memory anchors)

This is your final review map: a compact set of memory anchors tied directly to the course outcomes and to what AI-900 tests. Start by anchoring each domain to a “verb → service” pattern. For AI workloads: identify whether the problem is prediction (ML), perception (vision/speech), language understanding (NLP), or generation (Azure OpenAI/Copilot patterns). For ML fundamentals: remember supervised vs unsupervised vs reinforcement learning at a high level, plus classification vs regression vs clustering, and that Azure Machine Learning supports end-to-end lifecycle (data, training, deployment, monitoring).

For computer vision, lock in three separations: classification (label), object detection (label + location), OCR (text extraction). Remember the exam can test “responsible vision” principles: privacy considerations, bias risks, and when to avoid or limit certain vision use cases. For NLP workloads: sentiment, entities, key phrases, translation, speech, and question answering all map to distinct service capabilities; don’t collapse them into one generic “AI service.”

Exam Tip: Generative AI questions often test safe usage patterns more than model trivia. Think: grounding data sources (RAG), prompt instructions (system vs user intent), content filtering, and protecting sensitive data. If the scenario is enterprise copilots, prefer answers that mention security boundaries and responsible AI controls.

Create a one-page “last-minute sheet” with these anchors and your personal confusion pairs from the mocks. Your goal is fast recognition under pressure, not relearning.

Section 6.6: Exam day checklist—setup, timing, and anxiety control

On exam day, eliminate preventable failure modes first: environment, timing, and mental state. Ensure your test setup is stable (device charged, reliable internet, quiet space if remote). If the exam is proctored, comply with workspace requirements early to avoid time loss. Plan to start with a calm buffer so you’re not carrying stress into the first questions.

Timing: commit to the two-pass method you practiced. In the first pass, answer and move. In the second pass, apply elimination using service boundaries and constraints. If you feel stuck, write (mentally or on provided scratch space) the key verb and data type; then ask “which Azure service is designed for this exact verb?”

Exam Tip: Anxiety often shows up as rereading the same paragraph repeatedly. When that happens, jump to the question prompt (what are they asking?), then scan for constraint words (custom, real-time, private, generate, detect, extract). This breaks the loop.

Final checklist items: verify you understand the difference between prebuilt AI services and custom ML, confirm your generative AI safety vocabulary (content safety, grounding, access control), and remind yourself that AI-900 is a breadth exam. You do not need deep implementation details—your score comes from correct workload recognition and correct service selection. Finish by reviewing flagged answers only if you have time; don’t churn on already-solid choices.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
Chapter quiz

1. A company wants to build a chatbot that drafts responses to customer emails using the company’s product manuals. The manuals must not be used to train a public model, and access must be restricted using Azure identity. Which solution best meets the requirement?

Show answer
Correct answer: Use Azure OpenAI Service with Azure AD authentication and ground responses using your manuals (RAG) stored in your tenant
Azure OpenAI is the correct workload for generative text and can be secured with Azure AD and tenant-controlled data; grounding with your manuals aligns with common RAG patterns without training a public model. Computer Vision can extract/understand images and OCR text but does not provide an LLM for drafting natural-language email responses. Training a custom LLM from scratch in Azure Machine Learning is unnecessary for AI-900-level scenarios and is far more complex/costly than using a managed foundation model with grounding.

2. You are reviewing practice exam results. You consistently miss questions asking you to choose between Azure OpenAI, Azure AI Language, and Azure AI Vision. Which Weak Spot Analysis action is most effective?

Show answer
Correct answer: Create a one-page comparison matrix of each service’s primary workloads, typical inputs/outputs, and common distractors, then re-drill with timed question sets
A focused comparison matrix and targeted re-drills address the exam’s core skill: selecting the correct Azure AI workload/service and eliminating distractors. Pricing/SKU memorization is not a primary AI-900 objective and rarely helps with workload selection questions. Ignoring repeated misses prevents closing knowledge gaps and tends to repeat the same errors under exam pressure.

3. A team is building an app that must (1) identify objects in photos, (2) extract printed text from receipts, and (3) generate a short natural-language summary of the extracted text for the user. Which pairing of Azure services is the best fit?

Show answer
Correct answer: Azure AI Vision for object detection and OCR, and Azure OpenAI Service for summarization
Azure AI Vision covers computer vision workloads like object detection and OCR. Azure OpenAI is designed for generative tasks like summarization. Azure AI Language does NLP tasks (sentiment, key phrases, NER, etc.) and is not the right service for object detection/OCR. Azure OpenAI can describe images in some contexts, but for AI-900 service-boundary questions, dedicated Vision services are the expected choice for vision/OCR rather than treating an LLM endpoint as a replacement for vision capabilities.

4. During the exam, you notice a question with multiple requirements and tempting distractors. You have limited time remaining. Which approach best aligns with a repeatable AI-900 test-taking process?

Show answer
Correct answer: Identify the workload first (vision, NLP, generative AI, ML), map requirements to the correct Azure service boundaries, then eliminate options that violate security/responsible AI constraints
AI-900 commonly tests recognizing workloads and choosing the correct service while applying constraints like security and responsible AI; a structured elimination process reduces errors. Options with many product names are often distractors and can include unnecessary or incorrect services. Pure instinct without verification increases misreads and misses, especially in scenario-based questions.

5. A company wants to enable Microsoft Copilot features for employees. They are concerned about accidental exposure of sensitive internal data and want to reduce avoidable mistakes on rollout day. Which step is most appropriate to include in an Exam Day Checklist-style readiness checklist for this scenario?

Show answer
Correct answer: Verify tenant identity and access controls (e.g., least privilege), confirm data governance settings for where Copilot can pull content from, and review responsible AI guidance for acceptable use
Readiness should focus on identity/access (least privilege), data governance for content sources, and responsible AI usage policies—these align with security and responsible AI constraints emphasized in AI-900 scenarios. Disabling monitoring reduces visibility and weakens security/incident response. Granting global admin broadly violates least privilege and increases risk of misconfiguration and data exposure.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.