HELP

Microsoft AI Fundamentals for Non-Technical Pros (AI-900)

AI Certification Exam Prep — Beginner

Microsoft AI Fundamentals for Non-Technical Pros (AI-900)

Microsoft AI Fundamentals for Non-Technical Pros (AI-900)

Master AI-900 concepts and practice what Microsoft tests—without coding.

Beginner ai-900 · microsoft · azure · azure-ai-fundamentals

Prepare for Microsoft AI-900 with a non-technical, exam-first blueprint

This course is a complete, beginner-friendly prep path for the Microsoft AI-900: Azure AI Fundamentals certification exam. It’s designed for non-technical professionals who want to confidently describe AI concepts, recognize the right Azure AI services for common scenarios, and answer Microsoft-style questions under exam conditions—without requiring coding or data science experience.

What the AI-900 exam covers (official domains)

The AI-900 exam evaluates your understanding across these official domains:

  • Describe AI workloads
  • Fundamental principles of ML on Azure
  • Computer vision workloads on Azure
  • NLP workloads on Azure
  • Generative AI workloads on Azure

Throughout the course, you’ll practice selecting the right approach and service for realistic workplace scenarios—exactly the kind of decision-making the exam tests.

How this 6-chapter course is structured

The course is structured like a focused exam-prep book with six chapters:

  • Chapter 1 orients you to the AI-900 exam: registration options, scoring expectations, question types, and a study plan that works even if you’ve never taken a certification exam before.
  • Chapters 2–5 are domain-driven deep dives. Each chapter explains concepts in plain language, then reinforces them with exam-style scenario questions (service selection, terminology checks, and responsible AI decision points).
  • Chapter 6 is a full mock exam experience plus final review, weak-spot analysis, and an exam-day checklist.

Why this course helps you pass

Many AI-900 learners know the buzzwords but struggle with what the exam actually asks: choosing between AI workload types, interpreting basic ML concepts, and matching Azure services to business needs. This course closes that gap by:

  • Teaching the minimum effective technical depth needed for AI-900—no unnecessary math or coding
  • Emphasizing scenario-to-solution mapping (the most common exam pattern)
  • Including responsible AI considerations across domains, not as an afterthought
  • Providing a mock exam and review workflow to convert study time into points

Get started on Edu AI

If you’re ready to start preparing, set up your learning access and schedule your study plan. You can Register free to begin, or browse all courses to compare additional certification paths.

Outcome: exam-ready confidence for AI-900

By the end of this course, you’ll be able to describe AI workloads clearly, explain foundational ML principles on Azure, identify the right Azure services for computer vision and NLP scenarios, and articulate generative AI concepts and responsible usage—then prove it with a timed mock exam aligned to the official AI-900 domains.

What You Will Learn

  • Describe AI workloads and common AI solution scenarios, including responsible AI basics
  • Explain fundamental principles of machine learning on Azure and when to use Azure Machine Learning
  • Identify computer vision workloads on Azure and choose appropriate Azure AI Vision services
  • Identify natural language processing (NLP) workloads on Azure and choose appropriate Azure AI Language services
  • Describe generative AI workloads on Azure, including Azure OpenAI capabilities and responsible use considerations

Requirements

  • Basic IT literacy (browsers, cloud basics, and simple terminology)
  • No prior certification experience required
  • No programming or data science background needed
  • Ability to create or use a Microsoft account to explore Azure concepts (optional)

Chapter 1: AI-900 Exam Orientation and Study Strategy

  • Understand the AI-900 exam format, timing, and question styles
  • Register for the exam and set up your test environment (online or test center)
  • Build a realistic study plan and track your readiness by domain
  • Learn how to approach Microsoft-style questions and eliminate distractors
  • Set up a lightweight Azure learning environment (optional) and resource checklist

Chapter 2: Describe AI Workloads (AI-900 Domain Deep Dive)

  • Differentiate AI, machine learning, and deep learning in plain language
  • Match real business problems to AI workload types (vision, language, forecasting)
  • Explain key AI concepts: features, labels, training vs inference, and model drift
  • Apply responsible AI principles to common workplace scenarios
  • Complete domain-focused exam-style drills for 'Describe AI workloads'

Chapter 3: Fundamental Principles of Machine Learning on Azure

  • Identify regression, classification, clustering, and time-series use cases
  • Explain training/validation concepts and overfitting in non-technical terms
  • Choose between Automated ML, designer, and code-first workflows conceptually
  • Describe the Azure Machine Learning workspace and core components at a high level
  • Complete exam-style questions for 'Fundamental principles of ML on Azure'

Chapter 4: Computer Vision Workloads on Azure

  • Recognize when to use image classification, object detection, and OCR
  • Explain Azure AI Vision capabilities and typical business scenarios
  • Differentiate OCR/document reading and form-like extraction at a conceptual level
  • Understand facial analysis considerations and responsible use expectations
  • Complete exam-style practice for 'Computer vision workloads on Azure'

Chapter 5: NLP and Generative AI Workloads on Azure

  • Map business needs to NLP tasks: sentiment, key phrases, entities, summarization
  • Describe Azure AI Language capabilities and conversation/assistant use cases
  • Explain generative AI concepts: prompts, tokens, grounding, and safety
  • Describe Azure OpenAI use cases and responsible AI considerations for content generation
  • Complete mixed-domain practice for 'NLP workloads' and 'Generative AI workloads'

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
  • Final review sprint and confidence calibration

Jordan McAllister

Microsoft Certified Trainer (MCT) | Azure AI Fundamentals Specialist

Jordan McAllister is a Microsoft Certified Trainer who helps beginners pass Microsoft fundamentals exams through clear, business-friendly explanations. Jordan has supported learners preparing for Azure and AI certifications with focused practice aligned to official objectives.

Chapter 1: AI-900 Exam Orientation and Study Strategy

AI-900 is designed for non-technical professionals who need to speak confidently about AI concepts and the Azure services that implement them. This chapter orients you to what the exam measures, how to register and sit the exam, how Microsoft writes questions, and how to build a realistic plan that tracks your readiness by domain. You will also set expectations: AI-900 rewards accurate concept recognition and good service selection more than hands-on coding. That said, a lightweight Azure learning environment can be a force-multiplier—especially for understanding the difference between “what AI can do” and “which Azure product fits the scenario.”

Throughout this course, you will map every topic back to the outcomes: describing AI workloads and solution scenarios (including responsible AI), explaining machine learning fundamentals and when to use Azure Machine Learning, identifying computer vision and NLP workloads and the right Azure AI services, and describing generative AI workloads and Azure OpenAI capabilities responsibly. Start building the habit now: for any prompt, ask (1) what workload is this, (2) what service family fits, and (3) what responsible AI consideration should be mentioned.

Exam Tip: AI-900 questions often look like “business language.” Your advantage is translating business goals into a workload type (vision, language, ML, generative AI) and then choosing the most appropriate Azure service category—without overengineering.

Practice note for Understand the AI-900 exam format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set up your test environment (online or test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a realistic study plan and track your readiness by domain: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Learn how to approach Microsoft-style questions and eliminate distractors: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Set up a lightweight Azure learning environment (optional) and resource checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand the AI-900 exam format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set up your test environment (online or test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build a realistic study plan and track your readiness by domain: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Learn how to approach Microsoft-style questions and eliminate distractors: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: What AI-900 measures—official domains and weighting mindset

AI-900 tests foundational understanding rather than implementation details. The exam objectives are published by Microsoft and periodically updated, so your first step is to download the current skills outline and treat it as your checklist. The typical domains include: describing AI workloads and considerations (including responsible AI), describing fundamental principles of machine learning on Azure, and describing features of computer vision, natural language processing, and generative AI workloads on Azure.

Your “weighting mindset” matters even if you don’t memorize percentages. Allocate more study time to topics that produce many scenario questions: service selection (which Azure AI service fits), basic ML terminology (training vs. inference, classification vs. regression), and responsible AI principles. Lighter areas (like some process definitions) still matter, but they are usually tested as supporting details within scenario prompts rather than as standalone trivia.

What the exam is really measuring is whether you can: (1) recognize the workload (vision, language, generative AI, predictive ML), (2) pick the correct Azure service family (Azure AI Vision, Azure AI Language, Azure Machine Learning, Azure OpenAI), and (3) articulate basic considerations (data, accuracy, bias, privacy). Common traps include choosing a “powerful-sounding” option that is not the most direct fit, or confusing platform services (Azure Machine Learning) with prebuilt AI services (Azure AI services). Another trap is mixing up “model training” tools with “consuming a pretrained model.”

Exam Tip: When two answers look plausible, look for the one that most directly satisfies the scenario with the least custom work. AI-900 generally favors prebuilt services (Vision/Language/OpenAI) when the prompt describes common tasks like OCR, sentiment analysis, or chat-based summarization.

Section 1.2: Exam logistics—registration, pricing, ID, accommodations, policies

Logistics are easy to underestimate, and candidates lose points from avoidable stress. Register through Microsoft Learn or the certification dashboard and schedule with Microsoft’s exam delivery partner (often Pearson VUE). You can typically choose an online proctored exam or a test center appointment. Pricing varies by region and discounts (student, employer, event vouchers) may apply, so verify current cost before you commit.

For registration, ensure your legal name matches your government-issued ID. ID mismatch is a classic last-minute failure point. If you need accommodations, request them early through Microsoft’s accommodations process; approvals can take time and may require documentation. Understand policies: rescheduling windows, retake rules, and what counts as a violation during online proctoring.

Online testing requires a compliant environment: stable internet, a quiet room, desk cleared of papers, and often a webcam/microphone check. The proctor may ask you to show your desk and room; having a second monitor, phone, or notes nearby can trigger warnings or termination. For test centers, arrive early and plan for check-in time, lockers, and security procedures.

Exam Tip: Do a “test-run” of your online setup 24–48 hours before exam day: system check tool, webcam angle, lighting, and network stability. Treat environment issues as a study topic—because they can stop you from taking the exam at all.

Section 1.3: Scoring, passing, and question formats (MCQ, case, drag-and-drop)

AI-900 uses scaled scoring; the passing score is commonly published as 700 on a 1000-point scale, but the number of questions and exact scoring behavior can vary. Do not assume that every question is weighted equally. Some items may be unscored (used to evaluate future questions). Your goal is consistency across domains rather than perfection in one area.

Expect Microsoft-style formats such as multiple-choice (single answer), multiple response (“choose all that apply”), drag-and-drop matching (for concepts, services, or workflow steps), and scenario-based questions. Case study questions may present a longer business story, requirements, and then multiple questions tied to the same scenario. Read carefully: later questions may introduce new constraints (budget, compliance, latency) that change the best service choice.

Common traps include missing key words like “prebuilt,” “custom,” “real-time,” “on-device,” or “explainability.” Another frequent distractor is a correct statement that does not answer the question asked—for example, selecting Azure Machine Learning when the prompt clearly wants a prebuilt OCR feature (Azure AI Vision) or selecting a database/analytics service when the question is about model inference.

Exam Tip: For multiple-response items, don’t “average” your way to safety. Verify each selected option independently against the scenario requirements. A single wrong selection can cost you the whole item depending on scoring rules.

Time management is usually manageable, but don’t rush the first third of the exam. Establish a pacing rule: read the question stem, identify the workload and constraint, eliminate two distractors, then choose. If you are unsure, mark and move on—momentum prevents costly panic.

Section 1.4: Study strategy for beginners—spaced repetition and domain mapping

If you are new to AI and Azure, your biggest risk is “vocabulary overload.” The cure is a domain-mapped study plan with spaced repetition. Start by creating five buckets aligned to the outcomes: AI workloads/responsible AI, ML fundamentals/Azure Machine Learning, computer vision/Azure AI Vision, NLP/Azure AI Language, and generative AI/Azure OpenAI. Every study session should touch at least two buckets, and every week should revisit all five.

Spaced repetition means you re-encounter the same concepts repeatedly over increasing intervals (for example: Day 1, Day 3, Day 7, Day 14). This is particularly effective for service selection (which service does OCR? which does entity recognition? which is used for training vs. consumption?). Do not rely on a single long weekend of cramming—AI-900 questions are designed to test recognition under light ambiguity, which requires memory consolidation.

Build a simple readiness tracker. For each domain, rate your confidence (e.g., 1–5) on: definitions, service identification, and scenario selection. After practice, update your rating based on mistakes. Your study plan should follow data: if you consistently confuse Azure Machine Learning with Azure AI services, add a focused session on “custom ML vs. prebuilt AI.”

Exam Tip: Learn the “decision shortcut”: prebuilt service first (Vision/Language/OpenAI) unless the scenario explicitly requires custom model training, specialized datasets, or MLOps—then Azure Machine Learning becomes the likely answer.

Finally, consider setting up a lightweight Azure learning environment (optional). Even without coding, exploring the Azure portal and reading service descriptions makes the exam’s product names feel real, reducing cognitive load on test day.

Section 1.5: Responsible AI and ethics—how it appears across objectives

Responsible AI is not a separate “ethics-only” segment; it appears across scenarios in every domain. Microsoft commonly frames responsible AI around principles such as fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. On AI-900, you will not be asked to write policies, but you will be expected to recognize risks and select the best mitigation or best practice.

In ML contexts, responsible AI shows up as questions about bias in training data, explainability (being able to describe why a model made a prediction), and monitoring drift over time. In computer vision, risks can include demographic bias in facial analysis scenarios or privacy concerns with surveillance-like use cases. In NLP, watch for data leakage (sensitive customer text), toxic content, or misinterpretation of sentiment across cultures or languages. In generative AI, expect questions about groundedness (avoiding hallucinations), content filtering, prompt injection risks, and human oversight.

A common exam trap is choosing an answer that improves accuracy but ignores governance. For example, “collect more personal data” might increase performance but conflicts with privacy principles unless justified and protected. Another trap is assuming responsible AI is only about “being nice” rather than being systematic: documentation, access controls, auditability, and clear communication of limitations are all responsible AI actions.

Exam Tip: When you see words like “sensitive,” “regulated,” “public-facing,” “high impact,” or “customers,” mentally append a responsible AI checklist: privacy, bias, transparency, and human review. One of the options will usually align directly to these controls.

Section 1.6: Practice workflow—notes, flashcards, and review loops

Your practice workflow should be lightweight but disciplined. Start with a single page of “service-to-task” notes: map common tasks (OCR, image tagging, sentiment analysis, entity extraction, translation, chat completion, embeddings) to their Azure service families. Keep it simple and update it as you learn. The goal is fast recall under time pressure, not encyclopedic detail.

Use flashcards for high-yield distinctions: training vs. inference, supervised vs. unsupervised learning, classification vs. regression vs. clustering, and which scenarios call for Azure Machine Learning versus Azure AI services. Spaced repetition tools work well here, but paper cards are fine if you review them on a schedule.

Then apply a review loop: do a set of practice questions, tag every miss by domain and by mistake type (concept gap, misread stem, fell for distractor, or overthought). Your next study session should target the top two mistake types. This is how you “track readiness by domain” without guessing.

Exam Tip: When reviewing mistakes, rewrite the question in your own words as a workload statement: “This is NLP + sentiment analysis” or “This is vision + OCR.” If you can label the workload correctly, you can usually eliminate at least two options immediately.

Finally, if you set up an Azure learning environment, keep a resource checklist to avoid surprise costs: use free tiers where available, set budgets/alerts, delete test resources, and prefer guided sandboxes (like Microsoft Learn modules) when possible. The objective is familiarity, not building production solutions.

Chapter milestones
  • Understand the AI-900 exam format, timing, and question styles
  • Register for the exam and set up your test environment (online or test center)
  • Build a realistic study plan and track your readiness by domain
  • Learn how to approach Microsoft-style questions and eliminate distractors
  • Set up a lightweight Azure learning environment (optional) and resource checklist
Chapter quiz

1. You are advising a marketing manager who is preparing for AI-900. The manager asks what the exam is primarily designed to validate. Which statement best describes the focus of AI-900?

Show answer
Correct answer: Ability to recognize common AI workload types and select appropriate Azure AI service categories using business-oriented scenarios
AI-900 targets foundational understanding: mapping business needs to AI workloads (vision, NLP, ML, generative AI) and choosing appropriate Azure AI service families. Option B overstates hands-on coding and model-building depth, which is not the primary emphasis of AI-900. Option C aligns more with architect-level certifications; AI-900 does not test detailed production architecture decisions.

2. A company wants employees to take AI-900 from home. They ask what is most important to confirm when setting up the test environment for an online proctored exam. What should you emphasize first?

Show answer
Correct answer: Verify the testing space and device meet online proctoring requirements before exam day (quiet room, compatible system, and ID readiness)
For online delivery, meeting proctoring and system requirements is the critical first step to avoid being blocked from testing. Option B is unnecessary because AI-900 does not require SDK installation or multiple subscriptions. Option C is incorrect because certification exams do not allow using the Azure portal or searching documentation during the exam.

3. You are creating a study plan for AI-900 and want to track readiness effectively. Which approach best aligns with AI-900 preparation strategy described in the chapter?

Show answer
Correct answer: Track progress by exam domain areas and routinely test yourself with scenario questions to identify weak domains
AI-900 preparation is most effective when mapped to exam domains and reinforced with practice questions to reveal gaps by domain. Option B risks forgetting earlier content and does not provide readiness measurement across domains. Option C overemphasizes rote memorization; AI-900 favors correct workload recognition and appropriate service selection, not exhaustive SKU recall.

4. A question on the AI-900 exam describes this scenario: "A retailer wants to automatically categorize product photos and detect whether images contain a logo." What is the best first step to answer Microsoft-style questions like this?

Show answer
Correct answer: Identify the workload type (computer vision) and then choose the most appropriate Azure AI service family for that workload
Microsoft-style exam questions are often written in business language; the recommended approach is to translate the goal into a workload type (here, computer vision) and then pick the right Azure AI service category. Option B is a common distractor: not every scenario requires custom model training in Azure Machine Learning. Option C leads to overengineering; AI-900 expects the simplest appropriate service choice rather than the most complex.

5. A non-technical team wants a lightweight way to learn Azure AI concepts while preparing for AI-900. Which statement best describes the role of setting up a basic Azure learning environment for this exam?

Show answer
Correct answer: It is optional but can help reinforce the difference between what an AI workload is and which Azure product fits the scenario
A lightweight Azure environment is optional but useful for understanding how workloads map to services, which supports AI-900 scenario questions. Option B is incorrect because the exam does not include hands-on lab tasks. Option C is incorrect because practical exploration can improve conceptual clarity; the key is that it is not required, not that it is harmful.

Chapter 2: Describe AI Workloads (AI-900 Domain Deep Dive)

This chapter targets the AI-900 objective area that asks you to describe AI workloads and common AI solution scenarios. On the exam, you are rarely asked to build anything; you are asked to recognize the workload type (vision, language, forecasting, knowledge mining, or generative AI), understand basic machine learning (ML) terms, and apply responsible AI principles to realistic workplace contexts. Your job is to read a scenario, identify what the system must do (classify, predict, extract, generate), then map that to the right AI approach and Azure capability family.

A common candidate mistake is to treat “AI” as one tool. AI-900 expects you to differentiate AI vs ML vs deep learning in plain language: AI is the broad umbrella for systems that appear to perform intelligent tasks; machine learning is a subset where models learn patterns from data; deep learning is a subset of ML that uses multi-layer neural networks, often strong for vision, speech, and language. In exam scenarios, deep learning is usually implied (not named) when you see images, speech, or large-scale language understanding, but you still answer at the workload level unless the question explicitly asks about model types.

Throughout this chapter, focus on two exam habits: (1) translate business language into ML vocabulary (features, labels, training, inference), and (2) spot keywords that signal workload type (e.g., “detect objects in photos” → vision; “summarize emails” → generative AI + language; “search PDFs for answers” → knowledge mining). Exam Tip: When two options seem plausible, pick the one that matches the output the scenario needs (prediction, extraction, generation), not the input format.

Practice note for Differentiate AI, machine learning, and deep learning in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match real business problems to AI workload types (vision, language, forecasting): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain key AI concepts: features, labels, training vs inference, and model drift: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to common workplace scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Complete domain-focused exam-style drills for 'Describe AI workloads': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate AI, machine learning, and deep learning in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match real business problems to AI workload types (vision, language, forecasting): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain key AI concepts: features, labels, training vs inference, and model drift: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: AI workload categories—vision, language, knowledge mining, generative AI

AI-900 tests whether you can match a business problem to an AI workload category. Start by asking: “What kind of input do we have?” and “What kind of output do we need?” For vision workloads, the input is images or video and the output might be tags, detected objects, text extracted from images (OCR), or descriptions. Typical workplace examples include identifying damaged inventory from photos, reading license plates, or checking whether a factory worker is wearing safety gear.

Language workloads focus on understanding or producing meaning from text (and sometimes speech-to-text as a front door). Outputs include sentiment, key phrases, named entities (people, places, organizations), classification (routing support tickets), or translation. AI-900 often describes these as “natural language processing (NLP)” scenarios. Don’t confuse this with generative AI: traditional NLP typically extracts or classifies; generative AI produces new text.

Knowledge mining sits between content and search. The goal is to index and retrieve insights from large collections of documents—PDFs, scans, transcripts—so users can search, filter, and discover relationships. In practice, solutions use OCR plus language extraction and then store the enriched index for search. Exam Tip: If a scenario emphasizes “search across thousands of documents” or “build an index with extracted entities,” think knowledge mining rather than plain NLP.

Generative AI creates content: drafts, summaries, Q&A responses, or code-like text. It is best recognized by verbs like “generate,” “summarize,” “rewrite,” “compose,” or “chat.” On Azure, this points you toward Azure OpenAI capabilities. A classic exam trap is assuming generative AI is required whenever text is involved; if the task is to label, classify, or extract fields, traditional language services (or vision OCR) may be a better fit.

  • Vision: classify images, detect objects, OCR.
  • Language: sentiment, entities, classification, translation.
  • Knowledge mining: document ingestion → enrichment → searchable index.
  • Generative AI: new content (summaries, answers, drafts) with safety considerations.

Business mapping skill: “forecasting sales” is also an AI workload (prediction/regression) even though it isn’t vision or language. On AI-900, this typically appears as ML for numeric prediction (time series–like), not as generative AI.

Section 2.2: Core ML vocabulary—dataset, features, labels, training, inference

Core ML vocabulary appears constantly in AI-900 questions. A dataset is the collection of examples you use for learning or evaluation. A feature is an input attribute used to make a prediction (e.g., “account age,” “purchase history,” “image pixels,” “number of prior incidents”). A label is the correct answer for supervised learning (e.g., “fraud/not fraud,” “defective/not defective,” or a numeric value like “days to delivery”).

Training is when the algorithm learns patterns from a dataset. Inference is when the trained model is used to make predictions on new data. On the exam, training is associated with building or updating a model; inference is associated with deploying and using it in an app or workflow. Exam Tip: If the scenario says “use past data to build a model,” that’s training; if it says “predict for each new customer order,” that’s inference.

Understanding training vs inference helps you interpret Azure services questions. Azure Machine Learning is often referenced as the platform used to train, evaluate, and manage ML models. Many Azure AI services (Vision, Language) provide prebuilt models where you typically focus on inference, configuration, and responsible use rather than building the algorithm from scratch.

Model drift is another testable idea: performance can degrade over time because the real world changes (customer behavior shifts, new product categories appear, a camera’s lighting changes). Drift is not “the model forgot”; it is “the data distribution changed.” The practical response is monitoring, retraining, and validating with fresh data. A subtle trap: drift can happen even if your code is unchanged. If a scenario mentions “accuracy has dropped over the last quarter,” think drift and the need to review training data and model monitoring.

  • Supervised learning: uses labels; common for classification and regression.
  • Unsupervised learning: no labels; common for clustering and anomaly patterns.
  • Deep learning: an ML approach often used when features are high-dimensional (images, audio, language).

When you see “features” and “labels” in answer choices, you are being tested on your ability to map business columns to ML roles. A column can be a feature in one problem and a label in another depending on what you are trying to predict.

Section 2.3: Common metrics and outcomes—accuracy, precision/recall, error types

AI-900 expects you to speak in outcome terms: “How do we know the model is good enough?” The simplest metric is accuracy: the percentage of correct predictions. Accuracy is useful when classes are balanced (roughly equal numbers of each outcome). However, many business problems are imbalanced (fraud is rare, equipment failure is rare). In those cases, accuracy can be misleading: a model can be 99% accurate by always predicting “not fraud.”

That’s why you must recognize precision and recall. Precision answers: “When the model predicts positive, how often is it correct?” Recall answers: “Of all actual positives, how many did we catch?” These map directly to workplace risk: a medical screening system often prioritizes recall (don’t miss cases), while a high-cost investigation process may prioritize precision (don’t waste analyst time). Exam Tip: When a scenario emphasizes “minimize false alarms,” lean toward precision; when it emphasizes “don’t miss any real cases,” lean toward recall.

To reason about these metrics, you also need error types. A false positive is predicting something is true when it is not (flagging a legitimate transaction as fraud). A false negative is predicting something is false when it is true (missing a fraudulent transaction). The exam commonly frames this as “which error is more costly?” Your answer should align to business impact, not general ML theory.

For regression/forecasting (predicting a number), the exam may describe “difference between predicted and actual” as error. You don’t need advanced statistics, but you should recognize that lower error means better predictions, and that evaluation must happen on data not used for training (to avoid overestimating performance).

  • Accuracy: overall correctness; weak under class imbalance.
  • Precision: fewer false positives; “trust positive predictions.”
  • Recall: fewer false negatives; “catch most positives.”
  • Trade-off: tuning thresholds often shifts precision vs recall.

Common trap: thinking one metric is always best. AI-900 wants you to choose the metric that matches the scenario’s cost of mistakes.

Section 2.4: Responsible AI fundamentals—fairness, reliability, privacy, transparency

Responsible AI shows up as both direct questions and as “best choice” decision-making in scenarios. Four principles you must be able to apply in plain workplace terms are fairness, reliability and safety, privacy and security, and transparency.

Fairness is about avoiding unequal harmful outcomes across groups. For example, a hiring screening model trained on historical decisions could learn biased patterns and disadvantage certain applicants. The exam often tests whether you can identify this risk even when it is not explicitly called “bias.” Exam Tip: If the scenario involves people (hiring, lending, healthcare), expect fairness to be relevant and choose actions like evaluating performance across demographics and using representative data.

Reliability and safety means the system behaves consistently and fails gracefully. In AI terms, this includes robust performance under changing conditions and appropriate handling of uncertain predictions. If a model is used in a safety-critical environment, you should expect monitoring, human review, and fallback procedures. This pairs naturally with model drift: monitoring is both a reliability practice and a maintenance necessity.

Privacy and security is about protecting sensitive data used in training and inference. Scenarios might mention customer PII, medical records, or confidential documents. You are being tested on recognizing that AI solutions should follow least privilege, data minimization, and secure storage/processing practices. In generative AI scenarios, privacy includes preventing accidental leakage of sensitive prompts or outputs.

Transparency means stakeholders understand that AI is being used and can interpret outcomes appropriately. It includes explainability (“why did the model reject this loan?”), documentation, and communicating limitations. A frequent trap is choosing “make the model more complex” as a fix; transparency often improves when you add explanations, clear user messaging, and governance—not just accuracy.

  • Fairness: measure and reduce disparities across groups.
  • Reliability: monitor, test, and manage drift and failures.
  • Privacy: protect data; avoid unnecessary collection/exposure.
  • Transparency: disclose AI use; provide understandable rationale/limits.

When in doubt, pick the answer that adds monitoring, evaluation across groups, and clear communication—these are “safe defaults” aligned with AI-900 expectations.

Section 2.5: Selecting an AI approach—rules vs ML vs generative AI

AI-900 frequently asks you to select the most appropriate approach rather than a specific algorithm. Think of three buckets: rules-based logic, machine learning, and generative AI. Rules-based solutions are best when the logic is stable, explicit, and easy to write (e.g., “if invoice amount > $10,000 then require approval”). ML is best when patterns are too complex for explicit rules but you have historical data to learn from (fraud detection, demand forecasting, defect classification). Generative AI is best when you need flexible language generation or transformation (drafting, summarizing, conversational Q&A).

A key exam skill is identifying when ML is unnecessary. If a scenario is deterministic and unlikely to change, rules may be cheaper, easier to audit, and more transparent. Conversely, if the scenario includes “too many combinations,” “difficult to define rules,” or “needs to improve with data,” that points to ML. Exam Tip: Phrases like “learn from historical outcomes” and “predict” are ML signals; phrases like “compose,” “summarize,” and “rewrite” are generative AI signals.

Also distinguish prebuilt AI services vs custom ML. If the task is common and standardized (OCR, entity extraction, image tagging), prebuilt Azure AI services are usually the best match. If the task is unique to your organization (predicting churn from internal usage patterns), Azure Machine Learning is a typical fit for training custom models. The exam likes pragmatic choices: use prebuilt when it meets requirements; build custom when domain specificity demands it.

Generative AI introduces new considerations: factuality (hallucinations), policy compliance, and data protection. Even if generative AI can do a task, it may not be the best choice for structured extraction where deterministic output format matters. In those cases, traditional language extraction or a constrained ML classifier may outperform in reliability and auditability.

  • Rules: explicit, auditable logic; minimal data needs.
  • ML: prediction/classification from data; requires training, evaluation, monitoring.
  • Generative AI: produces new content; requires safety controls and human oversight for high-stakes use.

On AI-900, “best approach” answers usually balance capability with risk and operational simplicity.

Section 2.6: Practice set—scenario matching and concept checks (exam style)

This section prepares you for the domain-focused “Describe AI workloads” items without turning into a quiz. The exam pattern is consistent: you receive a short business scenario, then you must (1) identify the workload type, (2) identify whether it is training or inference, and (3) consider a responsible AI concern. Build a quick mental checklist you run in under 15 seconds.

Step 1: Identify the artifact. Images/video → vision. Text documents/emails/chats → language. “Search across documents and extract insights” → knowledge mining. “Generate a draft/summary/answer” → generative AI. Numeric forecasting (“predict next month’s demand”) → ML regression/forecasting.

Step 2: Identify the action verb. “Classify,” “detect,” “extract,” “translate,” “route,” “predict” usually indicate non-generative workloads. “Compose,” “summarize,” “rewrite,” “chat,” “create” usually indicate generative AI. Exam Tip: If you see “extract key phrases/entities,” do not pick generative AI just because text is involved—this is classic NLP extraction.

Step 3: Map to ML vocabulary. If the scenario mentions a historical outcome column (approved/denied, defective/not, dollars spent), that’s a label and implies supervised learning. If it says “group similar customers,” that’s clustering (unsupervised). If it says “use the model in an app to score new requests,” that’s inference. If it says “retrain monthly,” that’s training plus drift management.

Step 4: Choose the metric and watch for traps. If positives are rare and missing them is costly, recall matters. If false alarms are costly, precision matters. If the classes are balanced and the cost of each error is similar, accuracy may be sufficient. Be ready to justify error types: false positives vs false negatives, and which one the business can tolerate.

Step 5: Apply responsible AI. For people-impacting decisions, fairness and transparency are almost always relevant. For sensitive data, privacy and security are relevant. For changing environments, reliability and monitoring (including drift) are relevant. Many exam options include “increase training data” as a generic fix; prefer answers that directly address the stated risk (e.g., evaluate across demographics for fairness; add monitoring for drift; restrict access and protect PII for privacy).

  • Common trap: Selecting “AI” when rules suffice.
  • Common trap: Confusing knowledge mining (search + enrichment) with generic NLP.
  • Common trap: Using accuracy for imbalanced problems without considering precision/recall.
  • Common trap: Treating drift as a bug instead of a data change.

Use these steps to turn every scenario into a structured decision: workload category → approach (rules/ML/generative) → training vs inference → metric → responsible AI check. That is exactly what this AI-900 domain expects you to demonstrate.

Chapter milestones
  • Differentiate AI, machine learning, and deep learning in plain language
  • Match real business problems to AI workload types (vision, language, forecasting)
  • Explain key AI concepts: features, labels, training vs inference, and model drift
  • Apply responsible AI principles to common workplace scenarios
  • Complete domain-focused exam-style drills for 'Describe AI workloads'
Chapter quiz

1. A retail company wants to automatically detect whether product photos contain a visible barcode and, if present, draw a bounding box around it. Which AI workload type best fits this requirement?

Show answer
Correct answer: Computer vision (object detection)
This is a vision workload because the system must locate an object in an image (object detection with bounding boxes). Text analytics focuses on processing text, not pixels. Forecasting predicts future numeric values over time, which is unrelated to identifying barcodes in photos.

2. A manager says, "We want AI to approve or deny loan applications." You are translating the requirement into ML terms for a classification model. Which pairing correctly identifies the labels and features?

Show answer
Correct answer: Label: approval/denial outcome; Features: applicant attributes (income, credit score, debt-to-income)
In supervised classification, the label is the known target you want to predict (approve/deny) and the features are the input variables (income, credit score, etc.). Option B reverses the definitions. Option C misuses terms: drift describes changing data/model performance over time, and inference is the act of using a trained model to generate predictions.

3. A contact center wants an AI solution that can read incoming customer emails and generate a short summary plus a suggested reply draft for an agent to review. Which workload type is the best match?

Show answer
Correct answer: Generative AI (language)
Generating a new summary and draft response is content generation, which aligns with generative AI for language. Knowledge mining is primarily about extracting and indexing information from large document sets to enable search and insights, not drafting new replies. Forecasting predicts future values (like demand), not natural language summaries and responses.

4. A manufacturer trained a model last year to predict equipment failures from sensor readings. Recently, the model’s accuracy dropped because new machines were installed and the sensor patterns changed. What concept best describes this issue?

Show answer
Correct answer: Model drift
Model drift occurs when the statistical patterns in data change over time, reducing model performance. Training is the process of fitting a model to historical labeled data; it is not the phenomenon causing performance to degrade after deployment. Computer vision is a workload for images/video and is unrelated to sensor-pattern changes causing degraded predictions.

5. An HR team wants to use an AI model to screen resumes and rank candidates. Which action best supports the responsible AI principle of fairness?

Show answer
Correct answer: Evaluate model outcomes across demographic groups and mitigate biased patterns before deployment
Fairness requires assessing whether the model produces unequal outcomes for different groups and taking steps to reduce bias. Disabling explanations conflicts with transparency and accountability, not fairness. Never retraining can worsen harm over time due to drift and changing job requirements; responsible AI includes ongoing monitoring and improvement.

Chapter 3: Fundamental Principles of Machine Learning on Azure

This chapter maps directly to the AI-900 exam objective area that asks you to explain fundamental machine learning (ML) principles on Azure and when to use Azure Machine Learning. The exam is not looking for coding skill; it’s checking whether you can identify the right ML approach for a business problem, describe how models are trained and evaluated in plain language, and choose the right Azure Machine Learning workflow (Automated ML, designer, or code-first) at a conceptual level.

On the test, many questions are disguised as business scenarios: “predict next month’s demand,” “classify emails,” “group customers,” “detect unusual transactions.” Your job is to recognize the ML problem type and then connect it to how Azure Machine Learning helps you build, train, evaluate, and deploy a model.

As you read, keep two recurring exam patterns in mind: (1) the exam loves to test the difference between prediction (supervised learning like regression/classification) and pattern discovery (unsupervised learning like clustering), and (2) the exam frequently tests your ability to explain overfitting and why validation matters—without using heavy math.

Practice note for Identify regression, classification, clustering, and time-series use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain training/validation concepts and overfitting in non-technical terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose between Automated ML, designer, and code-first workflows conceptually: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Describe the Azure Machine Learning workspace and core components at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Complete exam-style questions for 'Fundamental principles of ML on Azure': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify regression, classification, clustering, and time-series use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain training/validation concepts and overfitting in non-technical terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose between Automated ML, designer, and code-first workflows conceptually: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Describe the Azure Machine Learning workspace and core components at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Complete exam-style questions for 'Fundamental principles of ML on Azure': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: ML problem types—classification, regression, clustering, forecasting

Section 3.1: ML problem types—classification, regression, clustering, forecasting

AI-900 expects you to identify common ML workloads and map them to the right problem type. The quickest way is to look at the output you want. If the output is a category, think classification. If the output is a number, think regression. If there are no labels and you want to discover natural groupings, think clustering. If the key requirement is predicting future values over time (often with seasonality), think forecasting (time-series).

Classification predicts which bucket an item belongs to: spam vs. not spam, loan approved vs. denied, “high/medium/low” risk. The exam often includes language like “assign,” “categorize,” “label,” or “detect whether.” Regression predicts a continuous value: house price, energy usage, delivery time in minutes. Look for “estimate,” “predict the amount,” “forecast a value” (but not necessarily time-series).

Clustering groups similar items without pre-defined labels: segment customers by purchasing behavior or group devices by telemetry patterns. You’ll see wording like “group,” “segment,” “discover patterns,” or “identify cohorts.” Forecasting is a specialized type of regression where time order matters: demand next week, call volume next hour, inventory next quarter. Exam scenarios may include seasonality, trends, or repeated time intervals.

  • Exam Tip: If the question mentions historical labeled examples (e.g., known outcomes), it’s usually supervised (classification/regression/forecasting). If it mentions no labels and “find groups,” it’s clustering.
  • Common trap: Confusing “forecasting” with “classification.” If the output is a numeric value over future time periods, it’s forecasting, not classification—even if the business talks about “high demand” informally.

When choosing between regression and forecasting on the exam, pay attention to time. Predicting “next month’s sales” implies time-series. Predicting “sales based on ad spend” could be regression if time isn’t central. The correct answer hinges on whether the ordering of observations (time) is essential to the model.

Section 3.2: Model training basics—splits, validation, generalization, overfitting

Section 3.2: Model training basics—splits, validation, generalization, overfitting

AI-900 frequently tests training and validation concepts using everyday language. The key idea: a model must perform well not only on the data it learned from, but also on new, unseen data. This is called generalization. To check generalization, we split data into at least a training set (what the model learns from) and a test set (what we hold back to evaluate final performance). Often, there is also a validation set used during model selection or tuning.

In non-technical terms, the training set is like practice questions, and the test set is like the final exam. If you only practice the exact same questions and memorize answers, you may score high in practice but fail on new questions. That’s overfitting: the model “memorizes” quirks and noise in training data instead of learning reliable patterns.

The exam also expects you to know that more complex models can overfit more easily, and that good evaluation needs data the model hasn’t seen. If a scenario says, “The model performs very well on training data but poorly in production,” that’s a classic overfitting clue.

  • Exam Tip: “Validation” is often described as “tuning the model” or “choosing the best version.” “Testing” is often described as “final evaluation” on held-out data.
  • Common trap: Training accuracy alone is not proof of a good model. If an answer choice celebrates high training performance without mentioning validation/testing, treat it as suspicious.

Another term you may see is underfitting, where the model is too simple to capture patterns, leading to poor performance even on training data. For AI-900, you typically just need to recognize overfitting vs. underfitting at a conceptual level and understand why data splitting and validation protect against misleading results.

Section 3.3: Feature engineering and data quality—why it matters for outcomes

Section 3.3: Feature engineering and data quality—why it matters for outcomes

Even though AI-900 is non-technical, it still tests the principle that models are only as good as the data and features provided. A feature is an input signal the model uses to make predictions—like age, location, purchase history, device temperature, or time of day. Feature engineering means choosing, transforming, or creating features to better represent the real-world problem.

Data quality issues show up on the exam as missing values, inconsistent formats, duplicate records, outliers, or biased/imbalanced data. These issues can cause inaccurate predictions or unfair outcomes. For example, if your dataset has far more “approved” than “denied” loans, a model may appear accurate by predicting “approved” too often, while failing the minority cases that matter most.

Exam Tip: When a scenario mentions “inconsistent data,” “different units,” “missing values,” or “not enough representative examples,” the right direction is usually data preparation/cleaning—not choosing a fancier algorithm.

  • Common trap: Assuming an Azure service automatically fixes bad data. Tools can help, but you still must address quality, relevance, and representativeness.
  • Common trap: Treating correlation as causation. A feature can be predictive without being a true cause; the exam may probe whether you understand this risk when interpreting models.

Feature engineering also includes encoding categories (e.g., product type), scaling numeric values, and extracting time-based signals (day of week, season) for forecasting. In AI-900 terms, you don’t need to know the math—just the reason: better features and cleaner data generally improve model performance and reliability.

Section 3.4: Azure Machine Learning overview—workspace, compute, datasets, models

Section 3.4: Azure Machine Learning overview—workspace, compute, datasets, models

Azure Machine Learning (Azure ML) is the primary Azure service for building, training, and deploying ML models. AI-900 expects you to recognize it as the “hub” for end-to-end ML work, especially when you need experiment tracking, model management, and deployment options.

The central container is the Azure Machine Learning workspace. Think of it as the project home where you manage assets and governance: experiments/runs, data connections, compute targets, models, endpoints, and monitoring integrations. If a question asks where you “organize and manage ML resources,” the workspace is usually the answer.

At a high level, Azure ML includes:

  • Compute: resources to run training jobs (compute instances/clusters). The exam won’t ask you to size CPUs, but it will test that training requires compute and can scale.
  • Data: datasets/data assets and connections to storage. You bring data in (or reference it) so training and scoring can use consistent sources.
  • Models: trained artifacts registered for versioning and deployment.
  • Endpoints: deployment targets for real-time or batch predictions (conceptually: “how apps call the model”).

Exam Tip: If the scenario mentions “track experiments,” “reproduce runs,” “register a model,” or “deploy a model as a service,” it’s pointing to Azure Machine Learning rather than a prebuilt cognitive service.

AI-900 also tests workflow choices conceptually. Azure ML supports a visual drag-and-drop designer, a guided Automated ML experience, and a code-first approach (SDK/CLI) for teams needing full control. The exam does not require syntax—only when you would choose each.

Section 3.5: Automated ML and responsible ML—model selection and interpretability basics

Section 3.5: Automated ML and responsible ML—model selection and interpretability basics

Automated ML (AutoML) in Azure ML helps you rapidly build a model by trying multiple algorithms and hyperparameters, then selecting the best candidate based on a metric (accuracy, AUC, RMSE, etc.). For AI-900, the core concept is: you supply labeled data and define the task (classification, regression, forecasting), and AutoML automates much of the experimentation.

AutoML is a strong fit when you want a good baseline quickly, have limited ML expertise, or need to compare models efficiently. In contrast, the designer is ideal when you want a visual pipeline for data prep and training without writing code, while code-first is best when you need custom logic, advanced control, or integration into software engineering workflows.

Responsible ML concepts show up in questions about transparency, fairness, and explainability. Even at a non-technical level, you should know the purpose of interpretability: helping humans understand which inputs influenced predictions and why. This supports trust, debugging, and compliance in regulated decisions (credit, hiring, healthcare).

  • Exam Tip: If the scenario asks for “the best model with minimal effort,” look for Automated ML. If it asks for “drag-and-drop pipeline,” look for designer. If it asks for “maximum flexibility/custom training,” look for code-first.
  • Common trap: Thinking AutoML removes the need for validation or responsible AI checks. AutoML automates experimentation, but you still must evaluate results, watch for bias, and validate on unseen data.

On AI-900, “responsible ML” is often tested as selecting actions that reduce harm: using representative datasets, monitoring performance drift, and being able to explain model decisions at a high level.

Section 3.6: Practice set—ML type selection and Azure ML concept questions

Section 3.6: Practice set—ML type selection and Azure ML concept questions

This section prepares you for the style of AI-900 questions without listing specific quiz items. Expect short scenarios where you must pick the ML type (classification, regression, clustering, forecasting) and then choose the right Azure ML concept (workspace, AutoML, designer, or code-first). The exam rewards fast recognition of keywords, but the safest method is to translate the scenario into “What is the output?” and “Do we have labels?”

For ML type selection, practice these mental checks: If the output is “yes/no” or a named category, it’s classification. If it’s a number, it’s regression—unless time is central, then it’s forecasting. If you are asked to “find segments” without outcomes provided, it’s clustering.

For Azure ML concept selection, look for cues: If the question emphasizes centralized management, governance, tracking runs, and deployments, it’s the workspace context. If it emphasizes trying multiple models automatically to find the best, it’s Automated ML. If it emphasizes visual steps and a no-code pipeline, it’s designer. If it emphasizes full control, custom training logic, or integration into development workflows, it’s code-first.

  • Exam Tip: Many wrong answers are “almost right” but mismatched to the scenario’s key constraint (time-series vs. regression, labeled vs. unlabeled, quick baseline vs. custom control). Circle the constraint first, then choose.
  • Common trap: Overcomplicating. AI-900 typically expects the simplest correct mapping based on the scenario, not an advanced ML architecture choice.

Use these checks to eliminate distractors quickly: any option that ignores data splitting/validation when asked about evaluation is likely wrong; any option that claims a model is “good” solely because training metrics are high is a red flag; and any option that treats clustering as a prediction of known labels is conceptually incorrect.

Chapter milestones
  • Identify regression, classification, clustering, and time-series use cases
  • Explain training/validation concepts and overfitting in non-technical terms
  • Choose between Automated ML, designer, and code-first workflows conceptually
  • Describe the Azure Machine Learning workspace and core components at a high level
  • Complete exam-style questions for 'Fundamental principles of ML on Azure'
Chapter quiz

1. A retail company wants to predict the total revenue for each store next month based on historical sales, promotions, and local events. Which machine learning approach is most appropriate?

Show answer
Correct answer: Regression
Regression is used for predicting a numeric value (next month’s revenue). Classification predicts a category label (for example, high/medium/low revenue bands) rather than an exact amount. Clustering groups similar stores without a known target value, so it does not directly predict revenue.

2. A bank wants to group customers into segments based on spending behavior and product usage, but it does not have predefined segment labels. Which type of machine learning should the bank use?

Show answer
Correct answer: Clustering
Clustering is an unsupervised learning technique used to discover groups in unlabeled data (customer segmentation). Classification requires labeled examples of each segment to train on. Time-series forecasting focuses on predicting future values over time (for example, next week’s balance), not discovering customer groups.

3. A team trains a model that performs extremely well on the training data but performs poorly on new, unseen data. In non-technical terms, what is the most likely issue, and what practice helps detect it?

Show answer
Correct answer: Overfitting; evaluate using a separate validation dataset
This describes overfitting: the model has effectively memorized training patterns and does not generalize. Using a separate validation dataset helps reveal this gap between training performance and real-world performance. Underfitting would typically show poor performance even on the training data. Clustering/classes are unrelated to the core issue of generalization and validation.

4. A non-technical analyst wants to build and compare several models in Azure Machine Learning with minimal code, while automatically handling algorithm selection and hyperparameter tuning. Which workflow should they choose?

Show answer
Correct answer: Automated ML
Automated ML is intended to help you quickly train and compare models with minimal coding, including trying different algorithms and tuning settings. Code-first (SDK) is best when you want full control and are comfortable writing code. Designer is a drag-and-drop approach, but it does not inherently focus on automated algorithm selection/tuning in the same way as Automated ML, and the statement that Designer requires writing code is incorrect.

5. You are reviewing an Azure Machine Learning workspace at a high level. Which statement best describes what the workspace provides?

Show answer
Correct answer: A central place to manage assets such as datasets, compute, experiments/runs, and registered models for ML projects
An Azure Machine Learning workspace is the central resource for organizing ML work: it helps manage datasets/data connections, compute targets, experiments and runs, model registration, and deployments. Real-time translation/sentiment APIs are typically part of Azure AI services, not the core purpose of an AML workspace. A storage account can hold data, but it does not automatically train models simply due to uploads without an orchestrated ML process.

Chapter 4: Computer Vision Workloads on Azure

Computer vision questions on AI-900 are rarely about algorithms; they are about selecting the correct workload type and mapping it to the right Azure service family. Expect scenario-based prompts that describe a business need (for example, “find damaged items in warehouse photos” or “extract text from scanned invoices”) and ask what kind of computer vision task it is, and which Azure capability fits. This chapter teaches you to recognize the patterns quickly: classification vs. detection vs. segmentation vs. OCR; general image analysis vs. document reading; and when “custom” becomes necessary.

The exam also tests that you understand boundaries and responsible use. Vision solutions often handle biometrics or sensitive imagery, so you must be able to identify when consent, transparency, and restricted use apply—especially around facial analysis. Keep your focus on the objective: choose appropriate Azure AI Vision services for common vision workloads, and explain the difference between OCR/document reading and form-like extraction at a conceptual level.

Exam Tip: When you see the words “label the whole image,” think classification. When you see “locate items” or “draw boxes,” think object detection. When you see “pixel-level outline,” think segmentation. When you see “extract printed/handwritten text,” think OCR/Read.

Practice note for Recognize when to use image classification, object detection, and OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain Azure AI Vision capabilities and typical business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate OCR/document reading and form-like extraction at a conceptual level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand facial analysis considerations and responsible use expectations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Complete exam-style practice for 'Computer vision workloads on Azure': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize when to use image classification, object detection, and OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain Azure AI Vision capabilities and typical business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate OCR/document reading and form-like extraction at a conceptual level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand facial analysis considerations and responsible use expectations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Vision workload types—classification, detection, segmentation, OCR

Section 4.1: Vision workload types—classification, detection, segmentation, OCR

AI-900 expects you to distinguish the major computer vision workload types based on the output required. In image classification, the model assigns one or more labels to the entire image (for example, “contains a cat” or “this is a product photo of shoes”). Classification is the right choice when location does not matter—only whether the image belongs to a category.

In object detection, the solution returns both a label and a location (typically a bounding box) for each detected object. Choose detection when the scenario includes counting items, locating hazards, finding inventory on shelves, or identifying where something appears in a frame (for example, “detect helmets on workers and mark their positions”).

Segmentation goes beyond boxes and returns a pixel-level mask—useful when you need precise outlines (for example, separating a background from a foreground object, or identifying the exact area of a defect on a surface). Segmentation is less common on AI-900 than classification/detection, but it appears as a conceptual option.

OCR (optical character recognition) is about extracting text from images: printed text, handwriting, signs in a photo, or text in scanned documents. On exam items, the clue is that the desired output is text, not “understanding of the image.”

  • Classification: label the image (no location)
  • Detection: label + location (bounding boxes)
  • Segmentation: label + pixel mask (precise shape)
  • OCR: extract characters/words/lines

Common trap: If a prompt says “identify all products in an image,” many learners jump to classification. But “all products” implies multiple objects and usually location/counting—object detection is the better match. Another trap is confusing OCR with “key-value extraction.” OCR returns text; structured field extraction is a document intelligence pattern (covered later in this chapter).

Section 4.2: Azure AI Vision overview—image analysis and OCR/Read concepts

Section 4.2: Azure AI Vision overview—image analysis and OCR/Read concepts

For AI-900, treat Azure AI Vision as the primary service family for general image understanding and text extraction from images. The exam often frames choices as “use a prebuilt capability” versus “train a custom model.” Azure AI Vision supports common tasks like tagging images, describing an image, detecting objects, and reading text (OCR). These capabilities are typically accessed through an endpoint and return JSON results (tags, captions, bounding boxes, or extracted text), which makes them easy to integrate into applications.

A critical concept is the difference between image analysis and Read/OCR. Image analysis focuses on visual content (what objects or scenes are present). Read/OCR focuses on transcribing text and often returns a hierarchy like pages, lines, and words with coordinates.

Exam Tip: If the scenario asks for “generate a caption,” “return tags,” or “detect common objects,” lean toward Azure AI Vision’s image analysis capabilities. If it asks to “extract text from photos or scanned pages,” select OCR/Read. Don’t overcomplicate with Azure Machine Learning unless the prompt says you need to build/train/manage your own model end-to-end.

Common trap: “Analyze a document” can mean two different things. If the goal is simply to read the text, Azure AI Vision Read fits. If the goal is to extract structured fields like invoice totals or line items, that’s typically a document extraction workload (conceptually different from plain OCR).

Section 4.3: Custom vision concepts—why and when customization is needed

Section 4.3: Custom vision concepts—why and when customization is needed

AI-900 expects you to know when prebuilt models are enough and when you should customize. Prebuilt vision models work well for broad, common concepts (everyday objects, generic scenes, common text). Customization becomes necessary when your categories are domain-specific, visually subtle, or not covered by general labels (for example, identifying a specific part number, a proprietary product variant, or distinguishing between similar defect types).

Customization also matters when your environment is unique: specialized camera angles, unusual lighting, medical or industrial imagery, or strict accuracy requirements. In these cases, you typically gather labeled examples and train a model for your specific classes (classification) or for locating your specific items (detection). On the exam, the signal words are “custom,” “company-specific,” “unique product,” “specialized,” or “not supported by built-in tags.”

Exam Tip: Look for the phrase “needs to recognize our own categories.” That is your cue that a custom vision approach is likely expected. Conversely, if the scenario is generic (tourism photos, standard retail items, common animals), prebuilt Vision is usually the intended answer.

Common trap: Learners assume “custom” automatically means “Azure Machine Learning.” For AI-900, customization might be described at a high level without requiring you to choose an ML platform. If the question is testing service selection for vision workloads, your best move is to choose the vision customization option when the requirement is domain-specific labels, not necessarily an end-to-end ML platform.

Section 4.4: Document and image text scenarios—receipts, IDs, invoices (conceptual fit)

Section 4.4: Document and image text scenarios—receipts, IDs, invoices (conceptual fit)

Text-in-image scenarios split into two conceptual levels that the exam loves to contrast. Level one is OCR/document reading: you want the words. This includes signs in photos, screenshots, scanned letters, or simple PDFs where returning the full text is enough for search, indexing, or accessibility.

Level two is form-like extraction: you want the meaning of the document’s structure—fields and values (for example, vendor name, invoice total, tax, dates) and sometimes repeating groups (like line items). Receipts, IDs, and invoices are classic prompts that hint you may need more than raw text. The exam does not require implementation details, but it does test whether you recognize that “extract total amount and merchant name from receipts” is not the same as “read the receipt text.”

  • Receipt scenario: often implies structured extraction (totals, dates) rather than just OCR.
  • ID scenario: may involve sensitive data; apply responsible AI thinking and minimal data handling.
  • Invoice scenario: commonly requires key-value pairs and tables/line items.

Exam Tip: If the prompt includes “key-value pairs,” “fields,” “tables,” “line items,” or “populate a database,” interpret it as form/document extraction. If it says “make the document searchable” or “extract all text,” interpret it as OCR/Read.

Common trap: Choosing image classification for document problems. Documents are rarely “classified by pixels”; they are interpreted by text and layout. When the asset is a receipt/invoice form, the right mental model is text + structure.

Section 4.5: Responsible vision AI—bias, consent, and sensitive use cases

Section 4.5: Responsible vision AI—bias, consent, and sensitive use cases

Responsible AI is part of AI-900, and vision workloads are where it becomes concrete. You must be able to explain why facial analysis and biometric-like scenarios require extra care. Key concerns include fairness (performance differences across demographic groups), privacy (images are personal data), transparency (people should know when vision AI is in use), and security (protect stored images and extracted data).

Facial analysis scenarios may involve detecting a face, analyzing attributes, or verifying identity. Even if the exam question is not technical, it may ask what consideration is most important. In those cases, think: consent, purpose limitation, and avoiding sensitive inferences. If a solution uses faces for access control, it must include user consent, strong governance, and fallback mechanisms.

Exam Tip: When you see “identify employees,” “track customers,” “detect emotion,” or “infer sensitive traits,” pause—these are high-risk prompts. The best answer often emphasizes responsible use (consent, privacy safeguards, bias evaluation) over raw capability.

Common trap: Treating “can the service do it?” as the only decision factor. On AI-900, the right answer may highlight that a use case is sensitive or requires strict controls, even if the technology exists. Also watch for answers that recommend collecting more personal data than needed; “data minimization” is a safe guiding principle.

Section 4.6: Practice set—service selection and scenario-based questions (exam style)

Section 4.6: Practice set—service selection and scenario-based questions (exam style)

AI-900 vision questions are primarily service-selection and workload-identification items. Your strategy is to translate the business wording into the output type, then map that output to the simplest matching Azure capability. Start by underlining what the system must return: a label, a location, a pixel mask, or text/fields.

Use this exam-style decision flow: if the output is text, choose OCR/Read. If the output is labels for an entire image, choose classification. If you need locations or counts, choose object detection. If you need precise shapes, choose segmentation. Then ask: are the categories generic (prebuilt Vision) or domain-specific (custom vision approach)? Finally, apply responsible AI checks if people, faces, IDs, or surveillance-like monitoring appear.

Exam Tip: Eliminate distractors by asking, “Does this choice produce the required output?” Many wrong options are plausible services but do not return the needed structure (for example, recommending image tagging when the scenario needs bounding boxes, or recommending OCR when the scenario needs object counting).

Common trap: Over-selecting complex platforms. If a prebuilt Azure AI Vision capability satisfies the requirement, that is usually the exam’s intended answer. Reserve Azure Machine Learning-style answers for prompts that explicitly say you must build, train, and manage custom models with your own data pipeline.

As you practice, also look for hidden constraints: “in real time” hints at fast inference; “on-device” might imply edge deployment; “regulated documents” hints at governance and privacy controls. AI-900 will not require deep architecture, but it will reward choosing the most appropriate, least-assumptive service for the described scenario.

Chapter milestones
  • Recognize when to use image classification, object detection, and OCR
  • Explain Azure AI Vision capabilities and typical business scenarios
  • Differentiate OCR/document reading and form-like extraction at a conceptual level
  • Understand facial analysis considerations and responsible use expectations
  • Complete exam-style practice for 'Computer vision workloads on Azure'
Chapter quiz

1. A retail company wants to automatically tag product photos as either "shoe", "shirt", or "hat" so that users can filter the catalog. The images contain only one main product per photo. Which computer vision workload best fits this requirement?

Show answer
Correct answer: Image classification
This is image classification because the goal is to assign a single label to the whole image (for example, shoe/shirt/hat). Object detection is used when you must locate multiple items and return bounding boxes, which the scenario does not require. OCR is for extracting text from images, not categorizing product types.

2. A warehouse team captures photos of pallets. They need to identify each box in the image and return its position so the system can count boxes and highlight missing ones. Which workload should you choose?

Show answer
Correct answer: Object detection
This is object detection because the requirement is to locate items in an image and return positions (typically bounding boxes) for each box. Image classification would only label the entire photo (for example, "pallet") and would not provide locations or counts. OCR would extract text (like labels on boxes) but would not reliably find and bound each box as objects.

3. A company scans invoices and wants to extract printed and handwritten text lines from the images so that the text can be searched and archived. Which Azure capability is the most appropriate starting point?

Show answer
Correct answer: Azure AI Vision OCR/Read
Use OCR/Read for extracting printed and handwritten text from images, which aligns with the exam objective to recognize OCR/document reading scenarios. Object detection and image classification are for finding/labeling visual objects, not extracting text content. While invoice processing may later require structured field extraction, the scenario explicitly asks for text extraction for search/archive, which is OCR.

4. A finance department receives many different vendor invoice layouts. They want to extract specific fields such as invoice number, total amount, and due date as structured key-value data, even when the text appears in different positions. Conceptually, which approach best matches this requirement?

Show answer
Correct answer: Form-like extraction (structured document field extraction)
This is form-like extraction because the goal is structured data (fields like invoice number/total/due date) rather than raw text. Basic OCR returns unstructured text and does not inherently map text to fields, especially across varying layouts. Image classification could label the document type (for routing) but would not extract the required values.

5. A gym wants to deploy cameras that identify specific members as they enter, without requiring badges. Which consideration best aligns with responsible AI expectations and typical Azure restrictions for facial analysis scenarios?

Show answer
Correct answer: Facial recognition is a sensitive/biometric use case and requires careful consideration such as consent and transparency; it may be restricted depending on the capability
Identifying specific people is a biometric facial recognition scenario and is treated as sensitive; responsible AI expectations include consent, transparency, and appropriate use, and some facial capabilities may be restricted. Treating it like ordinary object detection ignores the heightened privacy and compliance requirements. OCR cannot "read" identities from faces and does not avoid biometric concerns when the goal is identity recognition.

Chapter 5: NLP and Generative AI Workloads on Azure

This chapter targets the AI-900 objectives around Natural Language Processing (NLP) workloads and Generative AI workloads on Azure. On the exam, you are rarely asked to implement anything; instead, you are tested on recognizing the workload type from a business scenario and selecting the most appropriate Azure service family (Azure AI Language vs. Azure OpenAI, and when “classic NLP” is enough).

Expect scenario wording like “analyze customer feedback,” “extract company names,” “summarize call transcripts,” “build a chat experience,” or “generate marketing copy.” Your job is to map the business need to an NLP task (classification, extraction, summarization, translation) or to a generative task (content creation, Q&A with grounding), then identify the Azure product category that fits.

Exam Tip: When the requirement is to label or extract information from existing text (sentiment, key phrases, entities), think Azure AI Language. When the requirement is to create new text (draft, rewrite, brainstorm) or follow complex instructions conversationally, think Azure OpenAI—then add safety, grounding, and responsible use controls.

Practice note for Map business needs to NLP tasks: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Describe Azure AI Language capabilities and conversation/assistant use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts: prompts, tokens, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Describe Azure OpenAI use cases and responsible AI considerations for content generation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Complete mixed-domain practice for 'NLP workloads' and 'Generative AI workloads': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map business needs to NLP tasks: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Describe Azure AI Language capabilities and conversation/assistant use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts: prompts, tokens, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Describe Azure OpenAI use cases and responsible AI considerations for content generation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Complete mixed-domain practice for 'NLP workloads' and 'Generative AI workloads': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: NLP workload types—classification, extraction, summarization, translation

NLP workloads turn human language into structured signals or transformed text. AI-900 commonly tests four patterns: classification, extraction, summarization, and translation. Your first step in any scenario is to identify which of these the business wants.

Classification assigns a label to text. Typical labels include sentiment (positive/neutral/negative), topic categories, spam/not spam, or urgency levels. In business language, look for verbs like “categorize,” “route,” “detect tone,” or “flag.” Sentiment analysis is the classic example: analyzing product reviews to understand customer satisfaction.

Extraction pulls structured items out of text: key phrases, named entities (people, organizations, locations), or custom entities such as order numbers. Look for “extract,” “identify,” “pull out,” “tag names,” or “find mentioned products.” Key phrase extraction is not summarization—it returns short phrases, not a coherent paragraph.

Summarization reduces long text to a shorter representation, such as meeting notes, call center transcript summaries, or executive briefs. The exam may describe “too much text to read” or “provide a short overview.” Summarization can be extractive (selecting sentences) or abstractive (rewriting), but at fundamentals level you mainly need to recognize the use case.

Translation converts text from one language to another. Don’t confuse translation with transliteration (changing script) or with language detection (identifying the language). The scenario usually mentions “support multiple languages” or “translate support articles.”

  • Common trap: “Key phrases” vs “summary.” Key phrases are a list; summaries read like a short narrative.
  • Common trap: “Entities” vs “classification.” If the output is a set of items (names, dates), it’s extraction; if the output is a single label (category), it’s classification.

Exam Tip: If the scenario’s output can be stored cleanly in columns (SentimentScore, EntityName, EntityType), that’s a strong signal for classic NLP analysis rather than generative AI.

Section 5.2: Azure AI Language overview—analysis tasks and typical scenarios

Azure AI Language (part of Azure AI services) is the go-to choice on AI-900 for “analyze existing text” scenarios. It provides pre-built NLP capabilities that return structured results without you building a model from scratch. Typical exam scenarios include customer feedback analysis, document tagging, compliance monitoring, and insights extraction from emails, chats, or support tickets.

Map the business needs you see in prompts to these common Azure AI Language capabilities:

  • Sentiment analysis for understanding tone in reviews or social posts.
  • Key phrase extraction for identifying main topics mentioned in text.
  • Named entity recognition (NER) for pulling out people/organizations/locations; often used for indexing, redaction workflows, or analytics.
  • Entity linking (conceptually) for disambiguating entities (e.g., “Apple” the company vs fruit) in some solutions.
  • Summarization to shorten long text such as transcripts or articles.

The exam also expects you to recognize when a pre-built analysis service is more appropriate than training custom ML. If the requirement is common and general (sentiment, key phrases, standard entities), choose Azure AI Language rather than Azure Machine Learning.

Common trap: Overreaching to generative AI. If the question asks to “extract key phrases and entities” from thousands of documents, using a generative model is usually unnecessary and costlier; structured NLP analysis fits better and is easier to validate at scale.

Exam Tip: Watch for wording like “return a score,” “identify entities,” “extract,” “detect,” or “classify.” Those verbs align to Azure AI Language analysis outputs and are strong hints for the correct service selection.

Section 5.3: Conversational AI basics—bots, intent, utterances (concept-level)

Conversational AI on AI-900 is assessed at a concept level: you should understand what a bot is, what intents and utterances are, and why conversation scenarios differ from one-off text analysis. A bot is an application that interacts with users via chat or voice to answer questions or complete tasks (checking order status, resetting passwords, scheduling appointments).

An utterance is what a user says or types (for example, “Where is my shipment?”). An intent is the goal behind that utterance (for example, TrackOrder). Many different utterances can map to the same intent. Conversations may also capture entities (slots) such as order number, date, city, or product name to complete the task.

On the exam, conversational scenarios often include requirements like “maintain context,” “handle follow-up questions,” “route to human agent,” or “provide consistent answers.” In classic bot design, you typically detect intent and extract entities, then call back-end systems. In generative experiences, you may instead rely on an LLM to interpret the user request—but you still need guardrails and grounding (covered later).

  • Common trap: Thinking a chatbot is automatically generative. Many chatbots are retrieval/intent-based and do not generate free-form text beyond templates.
  • Common trap: Confusing “intent” with “sentiment.” Intent is the user’s goal; sentiment is their emotional tone.

Exam Tip: If the scenario emphasizes transactional outcomes (book, order, reset) and extracting structured fields, think “intent + entities” design. If it emphasizes drafting explanations or flexible language generation, that leans toward generative AI.

Section 5.4: Generative AI fundamentals—LLMs, prompts, temperature, hallucinations

Generative AI uses large language models (LLMs) to produce new content: text, summaries, answers, or transformations (rewrite, translate, classify via instructions). AI-900 expects you to know the core terms: prompts, tokens, temperature, grounding, and hallucinations.

A prompt is the input instruction and context you send to the model. Prompts can include role guidance (“You are a customer support agent”), constraints (“Respond in 3 bullet points”), and data (“Here is the product policy…”). Models process text in tokens (chunks of characters/words). Token limits matter because long prompts plus long responses can exceed the model’s context window—an exam-relevant concept when you see “very large documents” or “long chat history.”

Temperature controls randomness/creativity: lower values typically produce more deterministic, conservative outputs; higher values produce more varied outputs. On AI-900, don’t overthink numbers—just know the direction: lower for consistency (policy responses), higher for brainstorming (marketing ideas).

Hallucinations are plausible-sounding but incorrect outputs. They are not “bugs you can ignore”; they are a known risk of LLMs, especially when asked for facts beyond provided context. This is why grounding matters: connecting the model to trusted data (documents, databases) so responses are anchored in approved sources.

Common trap: Treating LLM output as guaranteed truth. If the scenario requires accuracy (regulatory, medical, financial), expect the best answer to include grounding and human review, not just “use an LLM to answer.”

Exam Tip: Look for keywords like “generate,” “draft,” “rewrite,” “create,” or “conversational responses.” Those are stronger indicators of generative AI than “extract” or “detect,” which usually indicate classic NLP analysis.

Section 5.5: Azure OpenAI on Azure—capabilities, deployments, and safe usage patterns

Azure OpenAI Service brings OpenAI models to Azure with enterprise features (identity, networking options, monitoring) and is positioned on AI-900 as the primary Azure service for generative AI workloads. Common capabilities you should recognize include generating text, summarizing, question answering, and assisting with content creation (emails, product descriptions) or conversational experiences.

A key concept is a deployment: in Azure OpenAI, you deploy a selected model and give it a deployment name your application calls. Exam items may describe “create a deployment,” “choose a model,” or “call the deployment endpoint.” You are not expected to code, but you should know that applications call deployments (not “the model” directly) and that different deployments can be tuned/configured for different scenarios.

Safe usage patterns are heavily emphasized in modern AI fundamentals. In scenario form, you may see: “prevent harmful content,” “avoid leaking sensitive data,” “ensure responses follow policy,” or “reduce hallucinations.” Strong answers usually include:

  • Content safety and filtering to detect or block disallowed content.
  • Grounding with approved enterprise data to reduce hallucinations and keep answers on-topic.
  • Prompting constraints (system instructions) to enforce tone, format, and refusal behavior.
  • Human-in-the-loop review for high-stakes outputs (legal/medical/finance).

Common trap: Assuming generative AI is the best default for any text task. If the goal is consistent structured extraction at scale, Azure AI Language is often more appropriate and easier to validate. Use Azure OpenAI when you need flexible language generation, instruction following, or natural conversational responses.

Exam Tip: If a question mentions “responsible AI,” “harmful content,” “jailbreaks/prompt injection,” or “need to cite sources,” that’s a cue to discuss safety controls and grounding rather than only model selection.

Section 5.6: Practice set—NLP vs GenAI selection and responsible AI scenarios

AI-900 “mixed-domain” questions often blend NLP and generative AI in one scenario to test your ability to pick the simplest correct approach. Your decision framework should be: (1) identify the desired output type, (2) decide whether structured analysis is sufficient, (3) if generating content, apply grounding and safety requirements.

Use these recognition patterns when you read a scenario:

  • If the output is a label/score (sentiment, category, priority), choose classification with Azure AI Language.
  • If the output is a list of items (names, organizations, key phrases), choose extraction with Azure AI Language.
  • If the output is a shorter version of existing content, summarization can be either classic NLP summarization or an LLM. Prefer Azure AI Language if it’s straightforward and needs consistent, auditable behavior; prefer Azure OpenAI if the summary must follow complex style instructions or combine reasoning with narrative generation.
  • If the output is new content (draft a response, rewrite with a tone, create FAQs), choose Azure OpenAI and apply safety/grounding.

Responsible AI scenarios typically test whether you know what can go wrong and what mitigations fit. If users may enter personal data, the safest answer includes data handling controls (minimize sensitive inputs), access control, and monitoring. If the model might fabricate details, include grounding to trusted sources and a review process. If the risk is harmful or biased content, include content filters and policy-aligned prompting.

Common trap: Answering with a service name but ignoring the constraint. For example, “use Azure OpenAI to answer employee policy questions” is incomplete if the question also states “must be accurate and only from the handbook.” The better exam answer couples Azure OpenAI with grounding to the handbook and safety controls.

Exam Tip: When two answers seem plausible, choose the one that (a) matches the output type and (b) explicitly addresses risks (hallucination, harmful content, data leakage) with practical mitigations.

Chapter milestones
  • Map business needs to NLP tasks: sentiment, key phrases, entities, summarization
  • Describe Azure AI Language capabilities and conversation/assistant use cases
  • Explain generative AI concepts: prompts, tokens, grounding, and safety
  • Describe Azure OpenAI use cases and responsible AI considerations for content generation
  • Complete mixed-domain practice for 'NLP workloads' and 'Generative AI workloads'
Chapter quiz

1. A retail company wants to automatically determine whether each customer review is positive, negative, or neutral and then trend the results over time. Which NLP task and Azure service family best fit this requirement?

Show answer
Correct answer: Sentiment analysis using Azure AI Language
The requirement is to label existing text with polarity (positive/negative/neutral), which is sentiment analysis—a classic NLP extraction/classification task supported by Azure AI Language. Azure OpenAI is primarily for generating new text or following complex instructions; it’s not the most appropriate first choice for simple sentiment labeling. Azure AI Speech is for converting audio to text; it does not perform sentiment classification by itself.

2. A legal team has thousands of contract paragraphs and needs to extract organization names and locations mentioned in the text to populate a database. Which capability should you use?

Show answer
Correct answer: Named entity recognition in Azure AI Language
Extracting organizations and locations from text is named entity recognition (entity extraction), which is provided by Azure AI Language. Summarization (whether via Azure AI Language or generative models) reduces text length but does not reliably return structured entity fields. Azure AI Vision analyzes images, not text documents.

3. A support center wants to provide agents with short summaries of long call transcripts to reduce after-call work. The transcripts are already available as text. What is the most appropriate approach?

Show answer
Correct answer: Use Azure AI Language text summarization to summarize the transcripts
The task is summarization of existing text, which is a classic NLP workload and aligns with Azure AI Language summarization features. Translation changes language, not length or key points. Sentiment classification does not provide a concise summary of the transcript content.

4. A marketing team wants an assistant that can generate first-draft product descriptions from bullet points and rewrite them in different tones. Which Azure service family is the best fit?

Show answer
Correct answer: Azure OpenAI
Generating new marketing copy and rewriting with tone/style constraints is a generative AI content-creation workload, best aligned with Azure OpenAI. Azure AI Language focuses on analyzing and extracting information from existing text (sentiment, entities, key phrases, summarization), not creative drafting. Azure AI Vision is for image/video analysis, not text generation.

5. A company is building a chatbot that answers employee questions using the company’s internal policy documents. They want responses to be based on those documents and to reduce the risk of the model making up facts. Which concept best addresses this requirement?

Show answer
Correct answer: Grounding the model with company data (e.g., retrieval-augmented generation) and applying safety controls
The requirement is to ensure answers are based on trusted internal sources and reduce hallucinations; grounding (often via retrieval of relevant documents) addresses this, and safety controls support responsible AI use. Increasing token limits affects how much text can be processed, but it does not ensure factual alignment to company policies. Key phrase extraction is an NLP analysis feature and does not create accurate, grounded answers to user questions.

Chapter 6: Full Mock Exam and Final Review

This chapter is your “dress rehearsal” for AI-900: you will simulate the real test, run two mixed-domain mock passes, diagnose weak spots, and finish with a tight review sprint. AI-900 rewards broad recognition more than deep engineering detail, so your goal is consistent service selection and clear mapping of scenarios to workloads: machine learning, computer vision, NLP, generative AI, and responsible AI. This chapter aligns directly to the exam outcomes: describing AI workloads and common solution scenarios; knowing when to use Azure Machine Learning; selecting Azure AI Vision and Azure AI Language capabilities; and understanding generative AI on Azure (Azure OpenAI) with responsible use considerations.

You will practice answering the way the exam expects: identify the workload first, then the Azure service family, then any key constraints (data type, real-time vs batch, custom vs prebuilt, governance, and safety). Many wrong answers on AI-900 are “near misses” from the right product family but wrong feature tier—your review clinic will train you to spot those distractor patterns.

Exam Tip: In AI-900, a correct answer often comes from one decisive keyword (e.g., “custom training,” “document extraction,” “speech,” “translation,” “chat completion,” “image tags,” “responsible AI”). Train yourself to circle that keyword mentally before you look at choices.

Use the sections below in order: simulate, attempt, attempt again, review deeply, remediate strategically, then lock in a calm exam-day routine.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final review sprint and confidence calibration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final review sprint and confidence calibration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam rules—timing strategy and how to simulate the test

To get value from a mock exam, you must simulate constraints, not just answer items casually. AI-900 is designed for breadth: you’ll shift rapidly among ML basics, Azure service selection, and responsible AI concepts. Your simulation rule: one sitting, timed, no notes, no searching. Use a quiet environment, a single screen, and a basic calculator only if your platform allows (most questions won’t require it).

Adopt a two-pass timing strategy. Pass 1: move quickly, answer what you know, and mark anything that needs rereading. Pass 2: return to marked items and force a decision using elimination. Do not spend disproportionate time on one confusing stem—AI-900 questions are intentionally short, and overthinking is a common failure mode.

Exam Tip: Your job is not to “prove” the answer; it’s to pick the best option given the scenario. If two options seem plausible, look for the constraint the exam writers want you to notice (custom vs prebuilt, structured vs unstructured data, online vs batch, or safety/governance requirements).

Simulate Microsoft-style question reading: first identify the workload category (prediction/classification/regression, vision, language, generative AI, responsible AI), then map to a service family (Azure Machine Learning, Azure AI Vision, Azure AI Language, Azure OpenAI). Only after that should you evaluate specific feature names. This prevents being pulled toward brand-recognition distractors.

Section 6.2: Mock Exam Part 1—mixed-domain questions and scenario sets

Mock Exam Part 1 should feel like a realistic first half: it mixes “what is AI?” conceptual items with scenario sets that require service matching. Expect rapid transitions: one item may test supervised vs unsupervised learning, the next may ask which Azure service extracts text from images, and the next may check responsible AI principles.

When you see a business scenario, translate it into inputs and outputs. Example mental pattern: “We have labeled outcomes and want predictions” → supervised ML. “We want to group similar customers without labels” → clustering (unsupervised). “We want to detect anomalies in sensor streams” → anomaly detection workload; then decide whether the question wants the general concept or a specific Azure offering.

For vision and NLP scenarios, anchor on the data type: images/video vs text/conversation. Then decide prebuilt vs custom. Prebuilt services are often the best choice when the scenario says “quickly,” “no ML expertise,” or “minimal training data.” Custom training is more likely when the stem mentions “domain-specific,” “company-specific,” “custom labels,” or “train a model.”

Exam Tip: Scenario sets frequently hide the key in one noun: “documents,” “receipts,” “forms,” or “invoices” usually signals document analysis/extraction rather than general OCR. Likewise “sentiment” and “key phrases” are classic NLP tasks; “tags” and “object detection” are classic vision tasks.

As you complete Part 1, track your confidence per item (high/medium/low). Confidence tracking is not fluff—it powers your weak spot analysis later so you can distinguish true gaps from careless reading.

Section 6.3: Mock Exam Part 2—mixed-domain questions and service selection

Mock Exam Part 2 should increase service-selection density: you will be asked to choose between closely related Azure options. This is where many non-technical candidates lose points—not because they don’t understand the workload, but because they confuse product families or choose a tool that is “capable” rather than “best fit.” Your discipline: pick the service that is purpose-built for the task described.

Machine learning selection: if the scenario emphasizes building, training, and managing models (datasets, experiments, pipelines, ML lifecycle), Azure Machine Learning is the safe anchor. If the scenario is simply “use AI without building models,” the exam often expects Azure AI services (Vision/Language) instead of AML. For generative AI, when the task is chat, summarization, or content generation with large language models, Azure OpenAI is the typical match.

Language scenarios often split into analysis vs conversation. “Extract entities, key phrases, sentiment” signals Azure AI Language. “Real-time speech-to-text or text-to-speech” is a different workload (speech), so do not let “language” wording trick you into choosing text analytics features when the input is audio. Vision scenarios similarly split: image understanding (tags, detection), text in images (OCR), and document extraction. Use the stem’s output requirement to decide.

Exam Tip: If the options include a general platform (like Azure Machine Learning) and a specialized cognitive service (Vision/Language/OpenAI), default to the specialized service unless the question explicitly requires custom model training, MLOps, or end-to-end ML lifecycle control.

Generative AI items often include responsible use constraints: you may need content filtering, prompt safety, or human oversight. Treat those as first-class requirements, not afterthoughts—AI-900 tests responsible AI basics as part of solution selection.

Section 6.4: Review clinic—explanations, distractor patterns, and common traps

Your review clinic is where you gain points fastest. Do not just mark answers right/wrong; write a one-sentence rationale tied to an exam objective: “This is computer vision (image input), so choose Azure AI Vision for image analysis,” or “This is supervised learning with labels, so it’s classification/regression.” If you cannot explain it in one sentence, you likely don’t own the concept yet.

Learn distractor patterns. Pattern 1: “Wrong family, right buzzword.” Example: choosing Azure Machine Learning for simple sentiment analysis because it sounds advanced. Pattern 2: “Right family, wrong feature.” Example: selecting generic OCR when the scenario wants structured field extraction from invoices. Pattern 3: “Overfitting to one keyword.” Example: seeing the word “language” and picking text analytics even though the input is speech audio. Pattern 4: “Ignoring constraints.” Example: choosing a custom model approach when the stem emphasizes minimal data or rapid deployment.

Exam Tip: If two answers look correct, ask: which one requires fewer assumptions? AI-900 favors the most direct, managed, purpose-built solution that matches the stated input/output and constraints.

Also review responsible AI traps. The exam may test fairness, reliability, privacy/security, inclusiveness, transparency, and accountability. A common trap is treating responsible AI as optional guidance rather than design requirements. If the stem mentions sensitive domains (HR, lending, healthcare), expect the correct answer to emphasize governance, explainability, privacy, and human oversight.

Finally, watch for “scope traps”: AI-900 is fundamentals. If an answer choice dives into deep infrastructure or niche engineering, it’s often a distractor unless the stem explicitly calls for it.

Section 6.5: Weak spot analysis—domain scoring, remediation plan, retake strategy

After both mock parts, score by domain, not just total. Create a simple grid: ML fundamentals, Azure Machine Learning use cases, computer vision, NLP, generative AI/Azure OpenAI, and responsible AI. For each missed item, label the failure type: (A) concept gap, (B) service confusion, (C) careless reading, or (D) overthinking. Your remediation plan depends on the type.

For concept gaps (A), revisit the workload definition and a single canonical example. For service confusion (B), build a “decision cue” list: what keyword pushes you toward Vision vs Language vs AML vs OpenAI. For careless reading (C), slow down on stems with negatives (e.g., “NOT,” “least likely”) and on questions that list multiple requirements. For overthinking (D), practice committing to the simplest managed service that meets requirements.

Exam Tip: Don’t remediate by rereading everything. Remediate by rewriting your own rules. Example: “If it’s unstructured text analysis → Azure AI Language; if it’s labeled prediction → Azure Machine Learning; if it’s image understanding → Azure AI Vision; if it’s content generation/summarization → Azure OpenAI.”

Plan a retake strategy even if you hope not to need it. If your mock score is borderline, schedule a short retake window (7–14 days) and focus on your lowest two domains first. Retaking without changing your study method usually reproduces the same mistakes; retaking after building decision rules and reviewing distractor patterns typically yields a quick improvement.

End this section by doing a “confidence calibration”: compare what you felt confident about versus what you actually got right. Your goal is to reduce false confidence (dangerous) and increase accurate confidence (stabilizing under time pressure).

Section 6.6: Exam day checklist—ID, environment, pacing, and final mental model

On exam day, eliminate avoidable stress. Confirm your ID requirements and name matching exactly. If testing online, validate your environment early: stable internet, power, quiet room, cleared desk, and allowed peripherals only. If testing in a center, arrive early enough to complete check-in without rushing your mindset.

Pacing: commit to the same two-pass method you used in the mocks. Pass 1 builds momentum and secures easy points; Pass 2 is where you use elimination and constraints. If you catch yourself rereading the same sentence repeatedly, mark it and move—this is a known sign of overinvestment.

Exam Tip: For every question, force the “final mental model” in five seconds: (1) What workload is this? (2) What data type is used? (3) Do we need custom training or prebuilt? (4) Which Azure service family fits? (5) Any responsible AI or governance constraint?

Use a final review sprint the day before: refresh your decision cues, revisit your most-missed domain, and rehearse responsible AI principles with one real-world example each. Avoid last-minute deep dives into advanced engineering topics; AI-900 rewards clear fundamentals and service recognition.

Finish with confidence calibration: you don’t need perfection—you need consistency. If you can reliably map scenarios to workloads and choose the most direct Azure AI service while respecting responsible AI basics, you are operating at the level this exam is designed to certify.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
  • Final review sprint and confidence calibration
Chapter quiz

1. A company wants a chatbot on its website that can answer questions using the company’s policy documents and must reduce the chance of unsafe or inappropriate responses. Which Azure service and feature set best fits this requirement?

Show answer
Correct answer: Azure OpenAI Service with chat completions + Azure AI Content Safety (and optional grounding via your documents)
This is a generative AI conversational scenario (chat completions). Azure OpenAI is the correct service family for chat-based generation, and responsible AI requirements point to using safety tooling such as Azure AI Content Safety (and grounding/RAG patterns). Azure AI Vision is for image/video workloads, not chat. Azure Machine Learning AutoML can train models but does not inherently provide a hosted chat LLM experience or built-in content safety for generated text in the same way.

2. You are reviewing a mock exam result and notice you frequently miss questions that involve extracting structured fields (invoice number, totals, vendor name) from scanned invoices. Which Azure AI capability should you choose in these scenarios?

Show answer
Correct answer: Azure AI Document Intelligence (prebuilt invoice models or custom extraction)
Extracting fields from documents such as invoices is a document processing workload. Azure AI Document Intelligence (formerly Form Recognizer) is designed for OCR + key-value/field extraction using prebuilt and custom models. Image tagging in Azure AI Vision is for labeling general image content (e.g., 'car', 'outdoor') and won’t reliably return invoice fields. Sentiment analysis is an NLP feature for evaluating opinion/emotion in text, not extracting document structure.

3. A retail chain wants to forecast weekly demand for each store. They have historical sales data in a table and want a guided approach that can automatically try multiple algorithms and select the best model. Which service should they use?

Show answer
Correct answer: Azure Machine Learning (Automated ML)
Demand forecasting from tabular historical data is a machine learning workload. Azure Machine Learning with Automated ML fits the requirement to automatically train and compare models. Azure AI Vision is for computer vision tasks like image classification/detection. Azure OpenAI focuses on generative language (and related) tasks rather than classical forecasting over structured datasets.

4. A global support team wants to translate customer emails from multiple languages into English in near real time. Which Azure service capability best matches this need?

Show answer
Correct answer: Azure AI Language - Translator
Language translation is a core NLP scenario covered by the Translator capability in Azure AI Language. OCR (Read) in Azure AI Vision extracts text from images/documents but does not translate it. Azure Machine Learning could theoretically build a custom translation model, but that is not the expected AI-900 service selection for a standard translation requirement.

5. On exam day, you see a question describing an app that must detect objects (e.g., 'person', 'forklift') in images from a warehouse camera feed. The solution should use a prebuilt model, not custom training. Which option should you pick?

Show answer
Correct answer: Azure AI Vision (prebuilt image analysis/object detection capabilities)
Detecting objects in images is a computer vision workload. If the requirement is prebuilt (no custom training), Azure AI Vision’s prebuilt image analysis/object detection capabilities are the best match. Azure Machine Learning implies building and training a custom model, which conflicts with the 'prebuilt' constraint. Azure AI Language entity recognition is for extracting entities from text, not identifying objects in images.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.