HELP

+40 722 606 166

messenger@eduailast.com

Microsoft Azure AI Fundamentals (AI-900) for Non-Technical Pros

AI Certification Exam Prep — Beginner

Microsoft Azure AI Fundamentals (AI-900) for Non-Technical Pros

Microsoft Azure AI Fundamentals (AI-900) for Non-Technical Pros

AI-900 made simple: understand Azure AI concepts and pass with confidence.

Beginner ai-900 · microsoft · azure · ai-fundamentals

Prepare for Microsoft AI-900 with a beginner-friendly, exam-aligned blueprint

This course is a complete exam-prep roadmap for the Microsoft Azure AI Fundamentals certification (AI-900). It’s designed for non-technical professionals who want to understand what AI can do, how Microsoft Azure positions AI services, and how to answer AI-900 questions confidently—even if you’ve never taken a certification exam before.

The AI-900 exam focuses on clear, practical fundamentals rather than coding. You’ll learn how to recognize AI workloads, explain core machine learning ideas, and map common business scenarios to the right Azure AI capabilities. Throughout the course, you’ll build a strong vocabulary for AI concepts and practice the decision-making patterns Microsoft tests.

Official exam domains covered (by name)

This course structure maps directly to the official AI-900 domains:

  • Describe AI workloads
  • Fundamental principles of ML on Azure
  • Computer vision workloads on Azure
  • NLP workloads on Azure
  • Generative AI workloads on Azure

How the 6-chapter book format helps you pass

Chapter 1 orients you to the exam: registration, typical question styles, scoring expectations, and a study strategy tailored for beginners. Chapters 2–5 each focus on one or two exam domains, using plain-language explanations that connect directly to real scenarios (for example: choosing between classification vs regression, when OCR is the right fit, or how to distinguish NLP analytics from generative AI creation). Each of these chapters includes exam-style practice to reinforce recognition and recall—key skills for fundamentals exams.

Chapter 6 then brings everything together with a full mock exam split into two parts, followed by targeted weak-spot analysis and a practical exam-day checklist. You’ll finish with a clear “last-mile” plan so your final review time goes to the topics that move your score the most.

What you’ll be able to do after this course

  • Explain AI concepts to stakeholders using Microsoft-aligned language
  • Identify which Azure AI workload (ML, vision, NLP, generative AI) fits a scenario
  • Understand responsible AI principles and how they appear in exam questions
  • Use practice sets and mock exams to steadily increase your readiness

Get started on Edu AI

If you’re ready to begin, create your learner profile and start the first chapter today: Register free. You can also explore other certification tracks and skill courses here: browse all courses.

By the end of this course, you’ll have a structured understanding of all AI-900 domains, a repeatable practice routine, and a mock-exam benchmark—so you can schedule your exam with confidence and pass on your first attempt.

What You Will Learn

  • Describe AI workloads: identify common AI scenarios, benefits, and responsible AI considerations
  • Explain fundamental principles of machine learning on Azure: training vs inference, features/labels, and model evaluation
  • Describe computer vision workloads on Azure: image classification, detection, OCR, and key Azure services
  • Describe NLP workloads on Azure: text analytics, conversational AI concepts, and Azure language services
  • Describe generative AI workloads on Azure: foundation models, prompt engineering basics, and Azure OpenAI concepts

Requirements

  • Basic IT literacy (web apps, cloud concepts, and using a browser)
  • No prior Microsoft certification experience required
  • No programming or data science background required
  • A computer with internet access to review Azure service concepts and practice questions

Chapter 1: AI-900 Exam Orientation and Study Game Plan

  • Understand the AI-900 exam format, question types, and scoring
  • Register for the AI-900 exam and plan your timeline
  • Set up an effective study routine for non-technical learners
  • Baseline diagnostic quiz and personal objectives
  • How to use Microsoft Learn and this course together

Chapter 2: Describe AI Workloads (Domain 1)

  • Recognize AI workloads vs traditional software solutions
  • Differentiate ML, computer vision, NLP, and generative AI scenarios
  • Apply responsible AI principles to business decisions
  • Exam-style practice set: AI workload identification
  • Scenario drill: choose the right AI approach

Chapter 3: Fundamental Principles of ML on Azure (Domain 2)

  • Understand core ML concepts: features, labels, training, and inference
  • Choose the right learning type: classification, regression, clustering
  • Interpret evaluation metrics at a fundamentals level
  • Know Azure ML concepts and AutoML at a high level
  • Exam-style practice set: ML fundamentals on Azure

Chapter 4: Computer Vision Workloads on Azure (Domain 3)

  • Identify vision tasks: classification, detection, and segmentation (fundamentals)
  • Understand OCR and document-intelligence scenarios
  • Match use cases to Azure vision services at a high level
  • Responsible AI and limitations in vision solutions
  • Exam-style practice set: computer vision on Azure

Chapter 5: NLP and Generative AI Workloads on Azure (Domains 4-5)

  • Identify NLP tasks: sentiment, key phrases, entities, summarization
  • Understand conversational AI concepts (bots and copilots) at fundamentals level
  • Explain generative AI basics: foundation models, tokens, and prompts
  • Apply prompt engineering patterns and safety concepts
  • Exam-style practice set: NLP + generative AI on Azure

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
  • Final review: domain-by-domain rapid recap

Jordan Mitchell

Microsoft Certified Trainer (MCT)

Jordan Mitchell is a Microsoft Certified Trainer who has helped learners prepare for Microsoft Fundamentals exams through role-based study plans and exam-style practice. Jordan specializes in translating Azure AI concepts for non-technical audiences while staying tightly aligned to Microsoft certification objectives.

Chapter 1: AI-900 Exam Orientation and Study Game Plan

The AI-900: Microsoft Azure AI Fundamentals exam is designed to confirm you can recognize common AI workloads and explain how Azure services support them—without requiring you to build models from scratch or write code. That makes it a strong entry point for non-technical professionals (project managers, business analysts, sales, procurement, compliance, and operations) who need to speak confidently about AI options, tradeoffs, and responsible use.

This opening chapter gives you an exam-first orientation: what the test is actually measuring, how to register and plan your timeline, how scoring and question styles work, and how to build a study routine that fits a non-technical schedule. You’ll also set expectations for how to use Microsoft Learn alongside this course so your practice mirrors the exam’s intent.

Exam Tip: AI-900 is less about memorizing service names and more about matching a scenario to the right AI workload (vision vs language vs ML vs generative AI) and describing concepts accurately (training vs inference, features vs labels, evaluation metrics, responsible AI principles).

Practice note for Understand the AI-900 exam format, question types, and scoring: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the AI-900 exam and plan your timeline: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Set up an effective study routine for non-technical learners: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Baseline diagnostic quiz and personal objectives: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How to use Microsoft Learn and this course together: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand the AI-900 exam format, question types, and scoring: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the AI-900 exam and plan your timeline: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Set up an effective study routine for non-technical learners: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Baseline diagnostic quiz and personal objectives: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for How to use Microsoft Learn and this course together: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: What AI-900 measures and who it’s for

AI-900 measures foundational understanding of AI concepts and Azure AI workloads. For non-technical candidates, the key is to think in “scenario-to-solution” terms: what problem is being solved (classify images, extract text, predict a value, summarize documents), what kind of data is involved (images, text, tabular rows), and what Azure capability typically supports it (Computer Vision, Azure AI Language, Azure AI services, Azure OpenAI, or Azure Machine Learning concepts).

The exam is aligned to broad outcomes: describing AI workloads and benefits, understanding machine learning basics (training vs inference, features/labels, evaluation), describing computer vision and NLP workloads, and recognizing core ideas behind generative AI and Azure OpenAI. You are not expected to tune models with code, implement algorithms, or configure deep infrastructure. Instead, you are expected to use correct vocabulary and make responsible recommendations.

Responsible AI considerations show up repeatedly. Expect questions where the “best” answer is the one that reduces harm: protecting privacy, avoiding bias, improving transparency, and ensuring human oversight. This is where non-technical professionals can score well—because the exam rewards thoughtful governance and risk awareness, not just tool selection.

Exam Tip: When two answers look plausible, the exam often differentiates by scope. Choose the option that best matches the workload category (vision vs language vs ML vs generative) before worrying about specific product names.

Section 1.2: Registration, scheduling, and exam policies

Registering for AI-900 typically happens through Microsoft’s certification portal and an exam delivery partner. Your first planning decision is timeline: pick a test date that creates urgency but still allows repetition and practice. Many non-technical learners do well with a 2–4 week plan if they study in small, consistent blocks (for example, 30–45 minutes on weekdays plus a longer review session on weekends).

Choose your delivery mode (online proctored vs test center) based on your environment. Online proctoring is convenient but strict: you need a quiet space, stable internet, and compliance with room and desk rules. Test centers reduce tech risk but add travel and scheduling constraints. Either way, read the exam policies early so you’re not surprised on exam day.

Build a “registration checklist” into your game plan: confirm your legal name matches your ID, validate your login and profile, and run any required system test for online delivery. If you use accommodations, start that process early. Put your exam appointment on your calendar and plan your final review window (the last 48 hours should emphasize recall and light practice, not heavy new content).

Exam Tip: Treat scheduling as a learning tool. A firm exam date is one of the strongest predictors of follow-through—especially for busy, non-technical professionals balancing work and study.

Section 1.3: Scoring, passing standard, and question formats

Microsoft exams use scaled scoring. Your goal is not to “get X questions right,” but to demonstrate competence across the measured skills. The passing standard is published by Microsoft as a scaled score, and some questions may carry different weight. Practically, this means you should avoid over-investing in a single topic while ignoring others; breadth matters.

Expect a mix of question formats such as multiple choice, multiple response (choose all that apply), matching, ordering/sequence, and scenario-based items. Scenario questions are common: you’ll be given a business problem and asked to choose the best AI workload, the most appropriate service category, or a concept that applies (for example, identifying whether a task is training or inference).

Non-technical candidates often stumble on “nearly correct” options that use inaccurate terms. For example, a prompt about predicting house prices is a machine learning regression scenario, not computer vision or NLP. A prompt about reading receipts is OCR (vision). A prompt about extracting sentiment from reviews is text analytics (NLP). A prompt about summarizing documents or generating new content is generative AI.

Exam Tip: Before reading answer choices, label the scenario in your own words: “This is classification,” “This is OCR,” or “This is text summarization.” Then pick the option that matches your label. This reduces the chance you’re led by distractors.

Section 1.4: Study strategy mapped to official exam domains

Your study routine should mirror the exam’s domains so you build coverage and confidence at the same time. Map your plan to the course outcomes and treat each as a “domain block” you revisit repeatedly: (1) AI workloads and responsible AI, (2) machine learning fundamentals, (3) computer vision workloads, (4) NLP workloads, and (5) generative AI concepts and Azure OpenAI basics.

A practical weekly structure for non-technical learners is: learn → summarize → apply. Start with Microsoft Learn modules for the domain, then use this course to translate definitions into exam-ready patterns (what the exam is really asking, what distractors look like, and how to spot the workload). Close each block with short practice and a written “one-page sheet” of key terms: features vs labels, training vs inference, precision/recall basics, classification vs detection vs OCR, entity recognition vs sentiment analysis, and what makes foundation models different.

Include a baseline diagnostic early in your timeline—not as a score to judge yourself, but as a map of where you are. Then set personal objectives like “I will correctly identify the workload category from a scenario in under 20 seconds” or “I will explain responsible AI principles and give an example risk.” Your objectives should be behavioral and measurable, not just “study more.”

Exam Tip: Use Microsoft Learn for authoritative definitions and this course for exam pattern recognition. The combination is powerful: Learn gives you correctness; exam coaching gives you speed and decision-making.

Section 1.5: Common pitfalls for beginners and how to avoid them

Beginners often confuse “AI workloads” with “products.” The exam is designed to test your ability to choose the right approach, not just recall a brand name. If you understand the workload first, the service choice becomes easier. A second common pitfall is mixing up related tasks: image classification (what is in the image) vs object detection (where it is) vs OCR (what text is written). Similarly, NLP includes tasks like sentiment analysis, language detection, and key phrase extraction, while conversational AI refers to bot experiences and dialog flow concepts.

Machine learning fundamentals can feel abstract, but the exam focuses on plain-language understanding: training is when the model learns patterns from labeled data; inference is when the trained model makes predictions on new data. Features are inputs; labels are the known answers during training. Model evaluation is about judging quality and tradeoffs—especially classification metrics (accuracy, precision/recall) at a conceptual level.

Responsible AI is another area where candidates answer too narrowly. The exam may frame risks as fairness, reliability/safety, privacy/security, inclusiveness, transparency, and accountability. A common trap is picking an answer that improves performance but ignores risk controls (for example, deploying without human review in a high-impact scenario).

Exam Tip: When responsible AI is mentioned, slow down and look for the option that adds governance: monitoring, human oversight, privacy protections, documentation, or bias evaluation—rather than “just use a better model.”

Section 1.6: Practice approach: eliminating answers and time management

Your practice goal is to become fast at identifying what a question is testing. Start by underlining (mentally) the nouns and verbs: “classify,” “detect,” “extract,” “summarize,” “predict,” “chat,” “translate.” Those words usually map directly to an AI workload category. Next, eliminate answers that clearly belong to a different data type (for example, an image problem answered with a language-only tool) or a different activity (training vs inference confusion).

Use a two-pass approach on practice sets. Pass one: answer confidently and move on. Pass two: return to the questions you marked and apply deeper reasoning. This trains you not to burn time early. For scenario questions, do not debate every detail—your job is to choose the best fit, not a perfect system design.

When eliminating options, watch for distractors that are “true statements” but not the best answer. Example pattern: one option correctly defines AI, another correctly defines machine learning, but only one matches the scenario described. Also watch for overly broad answers (“use AI to solve it”) versus specific workload matches (“use OCR to extract printed text”).

Exam Tip: If two options both sound reasonable, ask: “Which one directly accomplishes the verb in the scenario with the least extra assumptions?” Exams reward direct alignment more than ambitious architecture.

Finally, treat practice as feedback, not judgment. Track errors by category: workload confusion, terminology, or responsible AI. Then adjust your weekly plan: revisit Microsoft Learn for definitions, and use this course to rehearse scenario recognition until your first instinct is usually correct.

Chapter milestones
  • Understand the AI-900 exam format, question types, and scoring
  • Register for the AI-900 exam and plan your timeline
  • Set up an effective study routine for non-technical learners
  • Baseline diagnostic quiz and personal objectives
  • How to use Microsoft Learn and this course together
Chapter quiz

1. You are advising a non-technical stakeholder who is starting AI-900. They ask what the exam primarily measures. Which statement best describes the intent of AI-900?

Show answer
Correct answer: Your ability to recognize common AI workloads and describe how Azure services support them, without needing to build models or write code
AI-900 is a fundamentals exam focused on identifying AI workload types (vision, language, ML, generative AI) and describing related Azure capabilities and concepts. Option B aligns more with role-based, hands-on exams that expect implementation skills. Option C is beyond the scope of AI-900 and focuses on low-level model development and performance tuning rather than fundamentals and scenario matching.

2. A project manager is building an AI-900 study plan and wants to minimize cramming while ensuring they can adjust based on early performance. Which approach BEST aligns with an exam-first study game plan?

Show answer
Correct answer: Take a short baseline diagnostic quiz early, set personal objectives based on results, then schedule regular study blocks and targeted review
An exam-first plan emphasizes diagnosing gaps early and iterating: baseline assessment, clear objectives, and a consistent routine with focused review. Option B delays feedback until the end, increasing the risk of discovering gaps too late. Option C overemphasizes memorization; AI-900 questions commonly require matching scenarios to workloads and explaining concepts, not just recalling names.

3. A company wants a chatbot that answers questions based on internal policy documents. They are not asking you to name a specific service, only to classify the workload correctly for exam-style scenario matching. Which AI workload is MOST appropriate?

Show answer
Correct answer: Natural language processing (language workload)
A document-question-answering chatbot is primarily a language/NLP scenario (and may involve generative AI), so classifying it as a language workload is the best match. Computer vision (Option B) is for analyzing images/video. Time-series anomaly detection (Option C) is for spotting unusual patterns in chronological numeric data (e.g., sensor readings), not answering questions from text documents.

4. During exam prep, a learner confuses training and inference. In an AI-900 context, which statement correctly distinguishes them?

Show answer
Correct answer: Training is when a model learns patterns from labeled data; inference is when the trained model is used to make predictions on new data
AI-900 expects you to understand core ML concepts: training fits the model using historical (often labeled) data, and inference uses the trained model to score or predict on new inputs. Option B reverses the definitions. Option C is incorrect because the distinction is foundational and frequently tested in scenario questions.

5. A business analyst wants to use Microsoft Learn along with a third-party AI-900 course. Which strategy BEST ensures practice mirrors exam intent?

Show answer
Correct answer: Use Microsoft Learn modules to reinforce concepts and terminology, then validate understanding with scenario-based questions that map workloads to use cases
Microsoft Learn aligns closely with the exam skills measured and helps reinforce correct terminology and conceptual understanding; pairing it with scenario-based practice reflects how AI-900 tests workload identification and concept comprehension. Option B is wrong because AI-900 is not a coding/implementation exam. Option C is wrong because while terminology matters, AI-900 commonly tests applying concepts to scenarios rather than pure memorization.

Chapter 2: Describe AI Workloads (Domain 1)

Domain 1 of AI-900 is less about coding and more about “workload recognition.” The exam repeatedly tests whether you can look at a business scenario and correctly label the AI approach (or recognize when AI is unnecessary). In this chapter you’ll practice separating AI workloads from traditional software solutions, differentiating machine learning (ML), computer vision, natural language processing (NLP), and generative AI, and applying Responsible AI principles to real business choices.

You should also be ready to identify the right Azure service family at a high level. The exam does not require you to build models, but it does expect you to know what training vs. inference means, what features and labels are, how model evaluation is described, and which workload family (vision, language, speech, decision support) a problem belongs to.

As you read, keep a “scenario drill” mindset: when presented with a short prompt, immediately ask: (1) Is this rules-based or data-driven? (2) What is the input data type (text, image, audio, numbers)? (3) What is the output type (class, score, extracted text, generated text)? (4) What risk/Responsible AI constraint is implied?

Practice note for Recognize AI workloads vs traditional software solutions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate ML, computer vision, NLP, and generative AI scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to business decisions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam-style practice set: AI workload identification: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Scenario drill: choose the right AI approach: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Recognize AI workloads vs traditional software solutions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate ML, computer vision, NLP, and generative AI scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to business decisions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam-style practice set: AI workload identification: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Scenario drill: choose the right AI approach: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: AI workloads and common use cases (prediction, classification, automation)

Section 2.1: AI workloads and common use cases (prediction, classification, automation)

An AI workload is appropriate when you cannot reliably write a fixed set of rules to solve the problem, or when the solution must adapt as patterns change. On AI-900, the most common “AI workload identification” patterns are prediction, classification, and automation/decision support.

Prediction usually means estimating a numeric value (often called regression): forecast next month’s demand, estimate delivery time, predict energy usage, or score the likelihood of a customer churning. The exam often describes prediction indirectly as “forecast,” “estimate,” “probability,” or “risk score.”

Classification means choosing a category: spam vs. not spam, approve vs. deny, defect type A/B/C, positive/neutral/negative sentiment. A frequent trap is confusing classification with extraction; if the scenario is “pull out invoice number,” that’s not classification—it’s text extraction (OCR + NLP).

Automation is where AI augments decisions or triggers actions: route support tickets, prioritize leads, detect anomalies that prompt investigation, or recommend next best action. AI-900 likes scenarios where AI assists humans rather than fully replacing them—especially when Responsible AI concerns exist.

Exam Tip: If the scenario says “based on historical data” or “learn patterns,” think ML. If it says “if X then Y” with stable logic, think traditional rules/automation—not AI.

  • Training vs. inference: Training learns from data; inference uses the trained model to make predictions on new data. Many candidates miss that “running a model in production” is inference, not training.
  • Features and labels: Features are input variables (age, purchase history); labels are the correct answers during training (churned: yes/no).
  • Model evaluation: The exam expects basic awareness (accuracy, precision/recall, error rates). Don’t overcomplicate metrics—focus on whether the model meets the business goal and risk tolerance.

When you’re unsure whether AI is needed, ask: “Can a simple rules engine solve this consistently?” If yes, AI is often unnecessary or risky.

Section 2.2: Types of AI: ML vs rules-based, deep learning overview

Section 2.2: Types of AI: ML vs rules-based, deep learning overview

AI-900 frequently contrasts rules-based solutions with machine learning. Rules-based systems are deterministic: the same input always yields the same output because developers wrote explicit conditions. They work well when the domain is stable and the logic is clear (tax calculations, workflow approvals with fixed thresholds).

Machine learning is data-driven: instead of coding rules, you provide examples and the algorithm learns patterns. ML is a strong fit when (1) there are too many rules to write, (2) the “rules” are fuzzy (e.g., what counts as a “good” résumé), or (3) the environment changes (fraud patterns evolve).

Common exam trap: A scenario may mention “automation” and “AI” in the same description. Automation alone does not imply AI. If the task is a simple workflow (send an email when a form is submitted), it’s traditional automation. Look for ambiguity, pattern recognition, or learning from data to justify ML.

Within ML, you may see references to supervised learning (labeled data), unsupervised learning (find structure in unlabeled data, like clustering customers), and reinforcement learning (learn by trial and reward, less common in AI-900 scenarios but may appear conceptually as “optimize decisions over time”).

Deep learning is a subset of ML that uses multi-layer neural networks. It is especially effective for unstructured data like images, audio, and free-form text. On the exam, deep learning is often implied by tasks such as image recognition, speech transcription, or advanced language understanding. You typically won’t be asked to design network architectures; you only need to recognize that deep learning powers many modern vision/language solutions.

Exam Tip: If the input is unstructured (image, audio, natural language) and the task is complex (recognize objects, understand intent), default to “deep learning-based AI workload” rather than classic rules or basic ML.

Section 2.3: Key workload families: vision, language, speech, and decision support

Section 2.3: Key workload families: vision, language, speech, and decision support

Domain 1 expects you to quickly map a scenario to the correct workload family. A reliable method is to identify the data type first.

Computer vision workloads use images or video. Typical tasks include image classification (what is it?), object detection (where is it and what is it?), and OCR (what text is in the image?). A common exam confusion is between classification and detection: classification assigns a label to an entire image; detection finds bounding boxes and labels multiple items within the image. OCR is specifically about extracting printed or handwritten text from images (receipts, IDs, invoices).

Language (NLP) workloads use text. These include sentiment analysis, key phrase extraction, entity recognition (people, places, organizations), language detection, and text summarization. NLP also covers conversational scenarios like intent recognition and question answering. The exam often describes this as “analyze customer reviews,” “categorize support tickets,” or “extract topics from documents.”

Speech workloads use audio. Common tasks include speech-to-text (transcription), text-to-speech, translation, and speaker-related features. The trap here is assuming “chatbot” always means speech—many chatbots are text-only. Speech is specifically about audio input/output.

Decision support is where insights guide action: recommendations, anomaly detection, forecasting, or risk scoring. These are usually ML prediction/classification problems presented in business terms (optimize inventory, detect unusual transactions).

Exam Tip: In scenario drills, underline the input: “photo,” “scanned document,” “call recording,” “chat logs,” “transaction history.” The input type usually determines the workload family faster than the business goal does.

Section 2.4: Responsible AI: fairness, reliability, privacy, transparency, accountability

Section 2.4: Responsible AI: fairness, reliability, privacy, transparency, accountability

AI-900 treats Responsible AI as a first-class skill: you must recognize risks and choose safeguards. Microsoft’s common Responsible AI principles show up repeatedly: fairness, reliability & safety, privacy & security, transparency, and accountability.

Fairness means the system should not produce biased outcomes for protected groups. A classic business scenario is lending, hiring, insurance, or admissions. The exam may hint at fairness concerns by referencing sensitive attributes (gender, ethnicity, disability) or “unequal approval rates.”

Reliability and safety focus on consistent performance and avoiding harmful outcomes. Think of systems used in healthcare, transportation, or critical operations. A model that performs well in testing but fails in a new region or season is a reliability issue (data drift).

Privacy and security relate to protecting personal data, controlling access, and ensuring compliance. If a scenario involves medical records, IDs, voice recordings, or customer conversations, privacy considerations are implied.

Transparency means stakeholders should understand that AI is being used and, when appropriate, how decisions are made. In exams, transparency is often linked to explainability: “Why was my loan denied?”

Accountability means humans remain responsible for outcomes. The trap is choosing an answer that implies “the model decided, so no one is responsible.” In real deployments, you need governance, review processes, and clear ownership.

Exam Tip: When a scenario involves high-impact decisions (finance, employment, healthcare), look for answers that include human review, monitoring, bias evaluation, and clear escalation paths—not just “deploy the model.”

Responsible AI also connects to business decisions: sometimes the “right AI approach” is to limit automation, use decision support instead of auto-approval, or collect better data before deploying.

Section 2.5: Azure AI services overview and when to use them (high level)

Section 2.5: Azure AI services overview and when to use them (high level)

AI-900 expects high-level familiarity with Azure service categories so you can choose the right tool for a workload. You are not expected to memorize every SKU, but you should recognize the major families and the “build vs. buy” decision.

Azure AI Services (often described as prebuilt AI) are used when you want ready-to-consume capabilities without training your own model from scratch. Examples include Vision capabilities (image analysis, OCR), Language capabilities (sentiment, entities, summarization, Q&A), and Speech capabilities (transcription and synthesis). These are ideal when the task is common and general-purpose.

Azure Machine Learning is used when you need to train, tune, and manage custom ML models—especially for prediction and decision support based on your organization’s historical data. If the scenario mentions “use our proprietary data to build a model,” “experiment and compare models,” or “MLOps,” Azure Machine Learning is the expected direction.

Azure OpenAI aligns with generative AI workloads: text generation, summarization, chat experiences, and embeddings for search and retrieval-augmented generation. If the scenario emphasizes generating new content (draft emails, create product descriptions, conversational responses), generative AI is likely the match rather than classic NLP analytics.

Common exam trap: Mixing up NLP analytics with generative AI. “Extract key phrases” and “detect sentiment” are analytics; “write a response,” “compose,” or “generate a summary in a specific tone” signals generative AI.

Exam Tip: If the requirement is “no training required,” lean toward Azure AI Services. If it’s “custom prediction using our labeled data,” lean toward Azure Machine Learning. If it’s “generate or transform language,” lean toward Azure OpenAI.

Section 2.6: Domain 1 exam-style questions and rationale review

Section 2.6: Domain 1 exam-style questions and rationale review

This domain is assessed through short, practical prompts where you must identify the AI workload and sometimes the best Azure approach. Your goal is to justify the selection in one sentence: “Because the input is X and the output is Y, this is a Z workload.”

For the practice set and scenario drills in this chapter, review your answers using a rationale checklist rather than memorizing phrases:

  • Identify the input modality: image/video (vision), text (language), audio (speech), tabular history (ML prediction/decision support).
  • Identify the output form: category (classification), number/probability (prediction), bounding boxes (detection), extracted text (OCR), generated content (generative AI).
  • Decide rules vs. learning: if stable logic exists, rules-based may be best; if patterns must be learned, choose ML/AI.
  • Check Responsible AI signals: high-impact decisions imply fairness, transparency, and accountability requirements; personal data implies privacy/security.
  • Map to Azure at a high level: prebuilt AI Services vs. custom Azure Machine Learning vs. generative Azure OpenAI.

Common exam trap: Overfitting the answer to one keyword. For example, “document” could mean OCR (vision) if the input is a scanned image, or NLP if the input is already text. Always confirm whether the document is scanned (image) or digital text.

Exam Tip: When two answers seem plausible, choose the one that best matches the data type and the required output. The exam writers often include a “close but wrong” option that matches the business goal but not the modality (e.g., choosing NLP when the input is a photo of text).

By the end of this chapter, you should be able to look at a scenario and confidently label it as ML, vision, language, speech, or generative AI—and explain the choice while flagging any Responsible AI considerations that could change the recommended approach.

Chapter milestones
  • Recognize AI workloads vs traditional software solutions
  • Differentiate ML, computer vision, NLP, and generative AI scenarios
  • Apply responsible AI principles to business decisions
  • Exam-style practice set: AI workload identification
  • Scenario drill: choose the right AI approach
Chapter quiz

1. A retail company wants to automatically route customer emails to the correct department (Billing, Returns, Technical Support) based on the email body. Which AI workload is most appropriate?

Show answer
Correct answer: Natural language processing (text classification)
Routing emails by analyzing the meaning of text is an NLP classification workload. Computer vision is for images/video, not email text. Generative AI focuses on creating new text (for example, drafting a reply) rather than assigning an existing message to a category.

2. A logistics company wants to predict whether a delivery will be late based on historical shipment data that includes distance, weather, carrier, and day of week. The outcome is either Late or On time. What type of machine learning is this?

Show answer
Correct answer: Supervised learning (classification)
Because the training data includes a known outcome label (Late/On time), this is supervised learning. The output is a category, so it is classification. Unsupervised learning is used when you do not have labels and want to discover groupings. Computer vision is unrelated because the inputs are structured tabular features, not images.

3. A manufacturer wants to detect defects in product photos on an assembly line and draw bounding boxes around scratches and dents. Which AI workload best fits?

Show answer
Correct answer: Computer vision (object detection)
Detecting and locating visual defects with bounding boxes is a computer vision object detection scenario. NLP entity recognition extracts information from text, not images. Time-series forecasting predicts future numeric values over time and does not localize objects in images.

4. A company currently uses an if/then rules engine to approve expenses: if amount < $50 and category is Meals, then auto-approve. The company is considering replacing it with an AI model, but the policy is stable and rarely changes. What is the best recommendation for this scenario?

Show answer
Correct answer: Keep a traditional rules-based solution because the logic is explicit and stable
AI-900 emphasizes recognizing when AI is unnecessary. If requirements are clear, deterministic, and stable, traditional software rules are often more appropriate and easier to explain/audit. Unsupervised clustering would not directly implement approval decisions and provides no guaranteed mapping to policy outcomes. Generative AI creating policies introduces unnecessary risk and does not address the core approval logic requirement.

5. A bank wants to deploy an AI model to help decide whether to approve loans. Regulators require the bank to be able to explain key factors that influenced each decision and to monitor for unfair outcomes across demographics. Which Responsible AI principles are most directly involved?

Show answer
Correct answer: Fairness and transparency
Monitoring unfair outcomes maps to the fairness principle, and explaining key decision factors maps to transparency (and interpretability in practice). Privacy and inclusiveness may still matter, but they do not directly address the stated requirements about bias monitoring and decision explanations. Reliability/performance is important, but on its own it does not satisfy regulatory needs for fairness and explanation.

Chapter 3: Fundamental Principles of ML on Azure (Domain 2)

Domain 2 of AI-900 tests whether you can talk about machine learning (ML) in plain business terms while still using the correct technical vocabulary: features vs labels, training vs inference, and “how good is the model?” evaluation basics. You are not expected to build models in code, but you are expected to recognize common ML scenarios and choose the right learning type (classification, regression, clustering) and the right high-level Azure approach.

This chapter connects core ML concepts to the way questions are phrased on the exam. You’ll see the same patterns repeatedly: identify what the organization is predicting, determine whether labeled data exists, decide if the output is a category or a number, and then pick appropriate metrics and Azure ML tooling. Exam Tip: Many wrong answers are “almost right” but swap one key term (for example, calling clustering “classification,” or using accuracy for an imbalanced problem). Train yourself to spot that single mismatch.

By the end of this chapter, you should be comfortable translating a real-world scenario into: (1) features/labels, (2) learning type, (3) training vs inference, and (4) a small set of evaluation measures—then describing how Azure Machine Learning supports the workflow at a high level.

Practice note for Understand core ML concepts: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right learning type: classification, regression, clustering: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Interpret evaluation metrics at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Know Azure ML concepts and AutoML at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam-style practice set: ML fundamentals on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand core ML concepts: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose the right learning type: classification, regression, clustering: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Interpret evaluation metrics at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Know Azure ML concepts and AutoML at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: ML basics: datasets, features, labels, and algorithms

Machine learning starts with a dataset: a collection of examples (rows) described by attributes (columns). In exam language, the attributes used to make a prediction are features. The thing you want to predict is the label (also called the target). For example, “customer age, region, and past purchases” are features; “will churn: yes/no” is the label.

AI-900 commonly tests whether you can identify features and labels from a scenario description. Exam Tip: If the prompt says “predict,” “estimate,” “forecast,” or “classify,” the outcome described is the label. Everything else that could help make that decision is a feature. A common trap is treating an identifier (CustomerID, TicketNumber) as a useful feature; on the exam, IDs are typically not meaningful predictive features unless the scenario explicitly says they encode information.

An algorithm is the method used to learn patterns from the data (for example, a decision tree or linear regression). A model is the learned result—the artifact you use later to make predictions. In Azure terms, you train a model and then deploy it so applications can call it for predictions.

  • Features: inputs used for prediction (numbers, categories, text converted to numeric form, etc.).
  • Label: the correct answer during training (what the model learns to predict).
  • Model: learned mapping from features to label (or to groups in clustering).
  • Inference: using the trained model to make predictions on new data.

Keep your wording crisp: the exam expects “features/labels” rather than “inputs/outputs” when describing supervised ML. That vocabulary choice is often the difference between two close options.

Section 3.2: Supervised vs unsupervised learning and common scenarios

Choosing the learning type is a core AI-900 skill. The decision is usually driven by one question: do you have labeled examples? If you do, it’s typically supervised learning. If you don’t, and you’re trying to discover structure or groups, it’s unsupervised learning.

Supervised learning covers two high-frequency exam categories:

  • Classification: predict a category (fraud/not fraud, approve/deny, high/medium/low risk). Even if there are many categories, it’s still classification.
  • Regression: predict a numeric value (sales amount, temperature, delivery time). A classic trap is thinking “prediction” always means classification; on the exam, regression is the go-to when the output is a continuous number.

Unsupervised learning most often appears as:

  • Clustering: group similar items when you don’t have labels (customer segmentation, grouping support tickets by similarity, discovering product groupings). Exam Tip: Clustering does not predict a known label; it discovers groupings. If the scenario says “we don’t know the categories in advance,” clustering is a strong match.

Scenario cues matter. “Predict next month’s revenue” implies regression. “Identify whether an email is spam” implies classification. “Find natural groupings of customers based on behavior” implies clustering. When two answers seem plausible, anchor on the nature of the output: category vs number vs groups.

Section 3.3: Training, validation, testing, and avoiding overfitting

AI-900 expects a clear distinction between training (learning from historical data) and inference (using the trained model on new data). Training produces a model; inference consumes that model to output predictions. In Azure, inference commonly happens via a deployed endpoint or a batch scoring job.

To evaluate a model honestly, data is typically split into:

  • Training set: used to fit the model.
  • Validation set: used to tune settings and compare approaches while building.
  • Test set: held back until the end to estimate real-world performance.

Exam Tip: If you see “used to tune hyperparameters” or “choose between models,” think validation. If you see “final unbiased evaluation” or “never used during training,” think test.

A key risk is overfitting: the model performs very well on training data but poorly on new data because it memorized noise or overly specific patterns. Overfitting shows up on the exam as “high training accuracy but low test accuracy.” Ways to reduce overfitting at a fundamentals level include using more representative data, simplifying the model, and using validation techniques such as cross-validation (conceptually: repeated splits to ensure stability).

Another frequent concept is data leakage—when information from the future (or from the label itself) accidentally becomes a feature. This can make validation results look unrealistically good. On the exam, if a feature is effectively “the answer,” it’s a red flag. Example cue: “include ‘chargeback status’ to predict fraud,” where chargeback occurs after the transaction—this would inflate performance during training but fail in production.

Section 3.4: Metrics: accuracy, precision/recall, confusion matrix, RMSE (conceptual)

Evaluation metrics appear in AI-900 as conceptual matching: pick the metric that makes sense for the business risk. For classification, the most common building block is the confusion matrix, which counts correct and incorrect predictions by category (true positives, false positives, true negatives, false negatives). You don’t need to compute it in detail, but you do need to know what it represents: a table summarizing prediction outcomes.

Accuracy is the proportion of correct predictions overall. It’s easy to understand—and easy to misuse. Exam Tip: Accuracy becomes a trap with imbalanced data (for example, fraud is rare). A model that predicts “not fraud” all the time can have high accuracy but be useless.

Precision answers: “When the model predicts positive, how often is it correct?” This matters when false positives are costly (for example, flagging legitimate transactions as fraud and blocking customers). Recall answers: “Out of all actual positives, how many did we catch?” This matters when missing positives is costly (for example, failing to detect fraud or failing to detect a disease). Exam questions often describe the cost: if “missing a positive” is worse, lean toward recall; if “raising false alarms” is worse, lean toward precision.

For regression, a common metric is RMSE (root mean squared error), which represents typical prediction error in the same units as the label (for example, dollars or days). Lower RMSE indicates predictions closer to actual values. The exam may also describe it as “how far off predictions are on average,” but remember RMSE penalizes larger errors more strongly than small ones.

When you’re choosing between metrics in an answer set, look for the output type first: classification metrics (accuracy/precision/recall/confusion matrix) vs regression metrics (RMSE). That simple filter eliminates many distractors.

Section 3.5: Azure Machine Learning: workspace concepts, pipelines, and AutoML overview

Azure Machine Learning (Azure ML) is Microsoft’s primary service for building, training, and deploying ML models. AI-900 tests it at a high level: what it is used for, and the purpose of key concepts like workspaces, compute, and AutoML—not detailed configuration steps.

An Azure ML workspace is the top-level container that organizes ML assets: datasets, experiments/runs, models, endpoints, and connections to resources. Think of it as the “project home” for ML. Exam Tip: If the question is about “managing models, runs, and deployments in one place,” the workspace is usually the correct concept.

Compute (compute instances/clusters) provides processing power for training and batch scoring. Training can be expensive; one benefit of Azure is on-demand scaling rather than buying hardware. Azure ML also tracks experiments (sets of runs) so you can compare model performance and reproduce results.

Pipelines represent repeatable ML workflows—such as data preparation, training, and evaluation steps chained together. The exam angle is typically “automation and repeatability,” especially for operationalizing ML (often called MLOps at a conceptual level). If an option mentions “orchestration of steps” or “repeatable training process,” pipelines are a strong match.

Automated ML (AutoML) helps select algorithms and hyperparameters based on a chosen task (classification, regression, time-series forecasting in many contexts) and a target metric. AutoML is often positioned for teams that want to build effective models quickly without deep algorithm expertise. The common trap is overstating AutoML: it does not remove the need for good data, clear labels, and appropriate evaluation. It also doesn’t automatically solve responsible AI concerns; you still need to review data quality, bias, and appropriateness for the decision being made.

Section 3.6: Domain 2 exam-style questions and scenario mapping

Domain 2 questions are usually short scenarios with one or two crucial clues. Your job is to map the scenario to ML fundamentals: identify label, select learning type, decide whether you’re training or performing inference, and pick the metric or Azure ML concept that best fits.

Use a repeatable mental checklist:

  • 1) What is the outcome? If it’s a category, think classification; if it’s a number, think regression; if there is no known outcome and you want groupings, think clustering.
  • 2) Do we have labeled historical data? If yes, supervised; if no, unsupervised (clustering) is often correct.
  • 3) Are we building or using? Building/tuning is training/validation; calling a deployed model to score new items is inference.
  • 4) What does “good” mean? Accuracy for balanced classification, precision/recall when costs differ, RMSE for regression.
  • 5) Which Azure concept is being tested? Workspace (organize/manage), compute (run training), pipelines (repeatable workflow), AutoML (automated model selection/tuning).

Exam Tip: Watch for wording like “segment customers,” “group similar,” or “discover patterns.” Those cues point to clustering—even if the answer set tries to lure you toward classification. Conversely, phrases like “predict whether” or “classify as” usually imply supervised classification with labels.

Common traps include confusing model evaluation with model deployment (test vs inference), assuming accuracy is always the best metric, and mixing up Azure ML (custom model building) with other Azure AI services that provide prebuilt capabilities. In Domain 2, if the scenario is about creating and training an ML model on your data, Azure ML and AutoML are the default mental starting points.

Chapter milestones
  • Understand core ML concepts: features, labels, training, and inference
  • Choose the right learning type: classification, regression, clustering
  • Interpret evaluation metrics at a fundamentals level
  • Know Azure ML concepts and AutoML at a high level
  • Exam-style practice set: ML fundamentals on Azure
Chapter quiz

1. A retail company wants to predict whether a customer will churn (Yes/No) based on features such as number of support tickets, months as a customer, and last purchase date. Which machine learning type should you use?

Show answer
Correct answer: Classification
This is a supervised learning scenario with labeled outcomes (churn Yes/No), so classification is appropriate. Regression is used to predict a numeric value (for example, revenue). Clustering is unsupervised and groups data without pre-existing labels, so it would not directly predict churn.

2. You train a model that predicts house prices from features such as square footage and neighborhood. When the model is used to generate a price prediction for a new house listing, what phase of the ML workflow is being performed?

Show answer
Correct answer: Inference
Generating predictions for new, unseen data is inference. Training is when the model learns patterns from historical labeled data. Feature engineering is the process of creating/selecting input variables; it can occur before training and does not describe producing a prediction result.

3. A bank builds a model to detect fraudulent transactions. Only 0.2% of transactions are fraud. Which metric is typically more informative than accuracy for evaluating this model at a fundamentals level?

Show answer
Correct answer: Precision and recall
With highly imbalanced classes, accuracy can be misleading (a model predicting 'not fraud' for everything can be ~99.8% accurate). Precision and recall help evaluate performance on the minority class (fraud) and are commonly referenced for classification evaluation. R-squared and MAE are regression metrics, so they do not apply to a fraud/not-fraud classification scenario.

4. A marketing team has a dataset of customers with demographics and purchasing behavior but no existing segment labels. They want to group customers into similar segments for targeted campaigns. Which learning type best fits this requirement?

Show answer
Correct answer: Clustering
The goal is to discover groups in unlabeled data, which is clustering (unsupervised learning). Classification requires labeled categories to learn from (for example, known segment labels). Regression predicts a continuous numeric value and is not intended for grouping customers into segments.

5. A product manager wants to quickly compare multiple algorithms and feature transformations to predict customer satisfaction (High/Low) without writing code. Which Azure capability best supports this requirement at a high level?

Show answer
Correct answer: Automated ML (AutoML) in Azure Machine Learning
AutoML in Azure Machine Learning is designed to automate model selection and hyperparameter tuning and can be used with minimal code for common ML tasks such as classification. Azure AI Vision is focused on image-based scenarios (for example, object detection) and does not address tabular satisfaction prediction directly. Azure OpenAI Service focuses on generative AI and language models rather than automating traditional supervised ML model comparisons for structured data.

Chapter 4: Computer Vision Workloads on Azure (Domain 3)

Domain 3 of AI-900 expects you to recognize common computer vision tasks, describe what they do (at a non-technical level), and choose the right Azure service for the job. The exam is not looking for API syntax or deep model architecture. Instead, it tests whether you can map a scenario ("count people in a store camera feed," "extract text from invoices," "flag unsafe content in images") to the correct workload pattern and Azure capability.

This chapter focuses on the three “big buckets” of vision tasks—classification, detection, and segmentation—plus OCR and document intelligence. You’ll also learn how Azure AI Vision fits into these scenarios and where the edge cases are (for example, when document processing is better addressed by Azure AI Document Intelligence than by basic OCR). Finally, we’ll cover responsible AI considerations that frequently show up in exam wording: data quality, bias, privacy, and limitations.

Exam Tip: When a question includes words like “where is it?” or “how many?”, it’s almost never classification. Those cues point to detection (bounding boxes) or segmentation (pixel-level regions). When it includes “extract text,” think OCR or document intelligence rather than generic image analysis.

Practice note for Identify vision tasks: classification, detection, and segmentation (fundamentals): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand OCR and document-intelligence scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match use cases to Azure vision services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Responsible AI and limitations in vision solutions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam-style practice set: computer vision on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify vision tasks: classification, detection, and segmentation (fundamentals): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand OCR and document-intelligence scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Match use cases to Azure vision services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Responsible AI and limitations in vision solutions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam-style practice set: computer vision on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Vision workload patterns: image analysis and video analysis (conceptual)

Computer vision workloads generally fall into two patterns: image analysis and video analysis. Image analysis is “single frame” reasoning—classify a photo, detect objects in an image, read text, or tag visual features. Video analysis adds time as a factor: you may run the same image tasks on each frame, then aggregate results across frames (for example, tracking an object or counting unique people over time).

On AI-900, you won’t be asked to design a streaming pipeline, but you can be asked to identify the conceptual difference: video analysis implies repeated inference and often needs handling of motion, changing lighting, and continuity. That often increases false positives if the solution is not tuned for the environment.

Common scenario phrasing to watch for:

  • Image analysis cues: “photo,” “scan,” “uploaded image,” “catalog,” “snapshot.”
  • Video analysis cues: “CCTV,” “camera feed,” “live,” “frames,” “over time,” “tracking.”

Exam Tip: If a question describes “monitoring a store entrance” or “analyzing traffic footage,” it’s hinting at video analysis, but the underlying AI task might still be detection (finding cars) or classification (identifying vehicle types). Don’t confuse the media type (video) with the vision task (detection/classification/segmentation).

Common trap: assuming video automatically means “facial recognition.” AI-900 is careful about responsible use, and many organizations avoid face identification. Unless the scenario explicitly mentions identifying people by face (and that’s permitted), keep your answer aligned with generic tasks like object detection, counting, or OCR on frames.

Section 4.2: Image classification vs object detection vs segmentation (what to choose when)

The exam heavily emphasizes choosing between classification, object detection, and segmentation. The simplest way to differentiate them is by the type of output you need.

Image classification answers: “What is in this image?” It returns one label (or a set of labels) for the whole image, such as “cat” or “damaged product.” Use classification when location does not matter and there is typically one main subject or you only need a global decision (pass/fail, safe/unsafe).

Object detection answers: “Where are the objects, and how many?” It returns labels plus bounding boxes (rectangles). Use detection for counting items on a shelf, locating defects on a manufacturing line, or finding vehicles in a traffic image.

Segmentation answers: “Which pixels belong to which object/region?” It returns a mask (pixel-level outline). Use segmentation when precise shapes matter: measuring area of crop disease on leaves, separating background from a person, or identifying the exact boundary of a tumor region in medical imagery (conceptually).

Exam Tip: If the question says “highlight,” “outline,” “mask,” “separate foreground/background,” or “pixel-level,” it is pointing to segmentation. If it says “bounding box” or “coordinates,” that is object detection. If it says “label the image,” “categorize,” or “is this a…,” that is classification.

Common trap: confusing detection with segmentation. Detection can tell you “there is a crack here,” but segmentation is what you want when you must quantify the crack area or trace its exact shape. Another trap: selecting segmentation when the scenario only needs a yes/no decision; the exam tends to reward the simplest sufficient approach.

Section 4.3: Optical character recognition (OCR) and document processing scenarios

Optical character recognition (OCR) is a vision workload that extracts printed or handwritten text from images. On the exam, OCR is triggered by scenarios involving signs, receipts, labels, forms, IDs, invoices, and “scanned documents.” OCR output is typically the recognized text plus structural hints such as line/word positions.

However, many real business problems are not just “read text,” but “understand a document.” That’s the distinction between basic OCR and document processing (sometimes called document intelligence): extracting key-value pairs, tables, and fields (like invoice total, supplier name, and date). The exam expects you to understand that OCR alone may give you text, but document processing aims to convert documents into structured data suitable for workflows.

Exam Tip: When the scenario says “extract the total amount and invoice number” or “process forms into fields,” choose a document-intelligence approach, not just generic OCR. OCR is necessary but not sufficient for field extraction tasks.

Common traps in OCR questions include:

  • Assuming perfect accuracy: OCR quality drops with blur, skew, low contrast, unusual fonts, or glare.
  • Ignoring language/layout: Multi-language documents, rotated pages, and complex tables increase errors.
  • Forgetting privacy: OCR often processes sensitive personal or financial data; exam questions may hint at compliance needs.

To identify the best answer, focus on the final business output. If the output is “searchable text,” OCR is enough. If the output is “populate a database with fields,” you need document processing capabilities on top of OCR.

Section 4.4: Azure services overview: Azure AI Vision and related capabilities

For AI-900, you should recognize Azure AI Vision as the primary service for common image analysis capabilities. At a high level, Azure AI Vision supports tasks like tagging and describing images, detecting objects, and reading text (OCR). The exam usually frames this as “prebuilt” capabilities—meaning you can use trained models without building your own from scratch.

You should also be aware of related capabilities commonly associated with vision solutions:

  • Image analysis: Generate tags/captions, detect objects, identify visual features.
  • OCR (Read): Extract printed/handwritten text from images and documents.
  • Spatial understanding and content moderation concepts: Some scenarios ask about detecting unsafe or inappropriate content in images; interpret these as vision analysis needs, not NLP.

Exam Tip: AI-900 often rewards choosing a prebuilt service when the scenario does not require custom training. If a question never mentions “train,” “labeled images,” or “custom categories,” it’s likely aiming at Azure AI Vision’s out-of-the-box features.

Common trap: mixing “custom vision model training” concepts with the AI-900 requirement. AI-900 may mention custom models, but most questions are: “Which service can you use to detect objects/read text/describe images?” Keep the answer anchored to Azure AI Vision unless the scenario explicitly calls for specialized document field extraction (then document intelligence becomes a better match).

How to pick the right service in exam wording: map the scenario to the task first (classification/detection/segmentation/OCR), then map the task to a service family (Azure AI Vision for image analysis/OCR; document intelligence for structured documents).

Section 4.5: Designing vision solutions: data quality, bias, and privacy considerations

AI-900 expects “responsible AI” thinking even for non-technical roles: you should identify limitations and risks in vision solutions and propose practical mitigations. Vision systems are highly sensitive to data quality: lighting, camera angle, resolution, motion blur, and occlusion (objects partially hidden). These factors can change model performance dramatically even if the model is “good” in lab conditions.

Bias and fairness can show up when the training or evaluation data is not representative. For example, a product-detection system trained mostly on one packaging design may fail on a new design, or a safety-gear detector trained on one worksite may underperform in different environments. In exam scenarios, look for hints like “works well in one location but not another” or “higher error rates for a particular group,” which signal bias or dataset mismatch.

Privacy and security are frequent themes. Vision solutions may capture faces, license plates, addresses, or medical information. Even when the technical task is simple, the design must consider consent, data minimization, secure storage, retention policies, and access control.

Exam Tip: If an answer choice mentions “improve accuracy by collecting more representative data” or “reduce risk by anonymizing/redacting sensitive content,” these are often the most defensible responsible AI actions. Avoid answers that imply “the model will be objective” or “accuracy is guaranteed.”

Common traps:

  • Overclaiming: “Detects all objects in any lighting” is unrealistic; the exam penalizes absolute statements.
  • Ignoring human oversight: High-stakes uses (compliance, safety, HR decisions) should include review processes.
  • Mislabeling limitations as model failure: Sometimes the issue is poor camera placement or low-quality inputs, not the AI service.

To design robust solutions, align data collection with real operating conditions, run pilot testing, monitor drift (e.g., new packaging, seasonal lighting), and document known failure modes.

Section 4.6: Domain 3 exam-style questions with explanations

On Domain 3, the exam typically assesses your ability to do “scenario-to-task-to-service” mapping. The highest-value skill is reading the scenario carefully for output requirements (label vs location vs pixel mask vs extracted text) and then choosing the simplest sufficient Azure capability.

Here’s how to approach exam-style items without getting trapped by distractors:

  • Step 1—Identify the output: A single label implies classification; bounding boxes imply detection; pixel regions imply segmentation; extracted characters imply OCR; extracted fields/tables imply document processing.
  • Step 2—Look for training cues: Words like “custom categories,” “your own products,” “labeled images,” or “train a model” indicate custom model needs. If absent, prefer prebuilt Azure AI Vision capabilities.
  • Step 3—Watch for document wording: “Invoices,” “forms,” “key-value pairs,” “tables,” “fields,” and “automate data entry” are classic document intelligence cues.
  • Step 4—Apply responsible AI filters: If the scenario mentions people, cameras in public areas, or sensitive info, expect at least one option about consent, minimization, or bias evaluation.

Exam Tip: Distractor answers often “upgrade” the solution unnecessarily (for example, choosing segmentation when detection is enough, or choosing a custom training approach when prebuilt OCR solves it). AI-900 often tests whether you can avoid overengineering.

Also expect conceptual questions that test definitions rather than services: e.g., the difference between classification and detection, or what OCR produces. In those cases, choose the option that best matches the type of result (label, box, mask, text). If two answers both seem plausible, pick the one that matches the business requirement stated in the prompt, not an unstated future requirement.

Finally, keep your vocabulary sharp: “classification” is about categories; “detection” is about location/count; “segmentation” is about pixel-level separation; “OCR” is text extraction; “document processing” is structured extraction. That set of distinctions is one of the most frequently tested areas in Domain 3.

Chapter milestones
  • Identify vision tasks: classification, detection, and segmentation (fundamentals)
  • Understand OCR and document-intelligence scenarios
  • Match use cases to Azure vision services at a high level
  • Responsible AI and limitations in vision solutions
  • Exam-style practice set: computer vision on Azure
Chapter quiz

1. A retail company wants to use a ceiling-mounted camera to count how many people are currently in a store and draw boxes around each person in the video feed. Which computer vision task best matches this requirement?

Show answer
Correct answer: Object detection
Object detection identifies and locates objects by returning bounding boxes and counts, which matches the requirement to draw boxes around each person and determine how many are present. Image classification assigns a label to an entire image (for example, "store is crowded") but does not indicate where objects are. Image segmentation labels pixels (for example, exact person outlines) and is more detailed than needed when bounding boxes are sufficient.

2. A logistics company needs to process scanned invoices and extract structured fields such as invoice number, vendor name, and total amount into a database. Which Azure service is the best fit at a high level?

Show answer
Correct answer: Azure AI Document Intelligence
Azure AI Document Intelligence is designed for document-processing scenarios where you extract structured fields from forms like invoices and receipts. General OCR in Azure AI Vision focuses on reading text but does not inherently model document structure and key-value field extraction as well. Azure AI Content Safety is for detecting harmful content, not extracting text or fields from documents.

3. A manufacturing company has images of circuit boards and needs to identify defective solder areas by highlighting the exact pixels of the defect region for downstream measurement. Which vision task should you choose?

Show answer
Correct answer: Image segmentation
Image segmentation is used when you need pixel-level regions ("highlight the exact pixels") for an object or defect. Object detection returns bounding boxes, which are not precise enough for measuring the defect area. Image classification labels the entire image (for example, "defective" vs. "not defective") and does not provide location or shape information.

4. A social media platform wants to automatically flag user-uploaded images that contain adult or violent content. Which Azure capability best aligns with this requirement?

Show answer
Correct answer: Azure AI Content Safety
Azure AI Content Safety is intended to detect and moderate harmful or unsafe content (such as adult or violent imagery). Azure AI Document Intelligence is for extracting information from documents like forms and invoices, not for content moderation. Object detection locates objects with boxes, but it is not the primary Azure service for determining whether content is unsafe according to moderation categories.

5. A company is building a vision solution to identify employees entering a secure area. During responsible AI review, the team is told the model performs significantly worse for certain skin tones due to limited diversity in the training images. What is the best action to address this issue?

Show answer
Correct answer: Collect and use a more representative dataset and evaluate performance across demographic groups
Responsible AI guidance for vision solutions emphasizes data quality and bias mitigation: improving dataset representativeness and validating performance across groups directly addresses disparate performance. Higher camera resolution may improve clarity but does not fix biased training data or unequal performance across demographics. Changing from detection to classification changes the task output (location vs. label) and does not inherently resolve fairness issues caused by non-representative data.

Chapter 5: NLP and Generative AI Workloads on Azure (Domains 4-5)

Domains 4 and 5 on AI-900 focus on language: first, understanding and extracting meaning from text (NLP), and second, generating new content using foundation models (generative AI). The exam expects you to recognize common language tasks, match them to the right Azure services, and explain core concepts like intent, entities, prompts, tokens, and safety guardrails—without needing to build or train deep models.

This chapter is written to help you “pattern match” exam scenarios. When a question describes extracting insights from customer feedback, you should think text analytics (sentiment, key phrases, entities, summarization). When it describes a chat experience, you should think conversational AI (bots/copilots, orchestration, user experience). When it describes generating text or code, you should think generative AI and Azure OpenAI concepts, including responsible use.

Exam Tip: AI-900 frequently tests vocabulary. If you can quickly define utterance, intent, entity, token, and prompt, you can eliminate wrong choices fast—even if you’re not sure of every service name.

Practice note for Identify NLP tasks: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand conversational AI concepts (bots and copilots) at fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI basics: foundation models, tokens, and prompts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply prompt engineering patterns and safety concepts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam-style practice set: NLP + generative AI on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify NLP tasks: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand conversational AI concepts (bots and copilots) at fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI basics: foundation models, tokens, and prompts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply prompt engineering patterns and safety concepts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam-style practice set: NLP + generative AI on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: NLP fundamentals: utterances, intent, entities, and text analytics concepts

Section 5.1: NLP fundamentals: utterances, intent, entities, and text analytics concepts

NLP (natural language processing) covers workloads where the input is text (or speech converted to text) and the output is an insight or structured information. On AI-900, you are expected to recognize the “shape” of common NLP tasks: sentiment analysis (positive/negative/neutral), key phrase extraction (main topics), named entity recognition (people, places, organizations), language detection, and summarization (condensing long content). These tasks are often grouped under “text analytics.”

For conversational scenarios, the exam uses three foundational terms: an utterance is what a user says (“Book me a flight to Seattle tomorrow”). The intent is what they want to do (BookFlight). The entities are the important variables inside the utterance (destination=Seattle, date=tomorrow). Distinguish entities from key phrases: entities are usually typed/structured categories (Person, Location, DateTime), while key phrases are salient words/phrases that summarize the content.

Summarization appears in two exam-friendly forms: extractive (selecting key sentences from the original text) and abstractive (generating a new condensed version). You don’t need implementation details, but you should know that summarization is an NLP workload and that “generate a new summary” leans toward generative approaches, while “pull important sentences” matches extractive summarization.

Exam Tip: When a scenario says “identify the main themes in a set of reviews,” prefer key phrase extraction. When it says “identify people, product names, and locations,” prefer entity recognition. When it says “determine how customers feel,” prefer sentiment analysis. These are easy points if you map verbs to tasks.

  • Sentiment: classify opinion polarity in text.
  • Key phrases: extract topic-defining phrases.
  • Entities: detect and categorize named items (and often link them to known entries).
  • Summarization: shorten content while preserving meaning.

Common trap: confusing “translation” with “sentiment” or “language detection.” If the question’s goal is understanding meaning in the same language, it’s text analytics; if it’s converting between languages, that’s translation (also language-related, but a different task category). Another trap is overthinking “training.” Many NLP insights can be obtained via prebuilt models (no model training required), which is central to the “fundamentals” level of AI-900.

Section 5.2: Azure language services overview: analyzing and extracting insights from text

Section 5.2: Azure language services overview: analyzing and extracting insights from text

On the exam, Azure’s language capabilities are usually presented as managed services you can call via APIs/SDKs to analyze text. Your job is to connect a business requirement to the correct service category. If a scenario describes mining insights from emails, tickets, or survey comments, think “Azure AI Language” capabilities for text analytics such as sentiment, key phrases, entity recognition, and summarization. These are positioned as prebuilt AI—quick to adopt and appropriate when you don’t need custom model training.

Expect questions that test the distinction between generic extraction (sentiment, entities) and custom classification or extraction. If a company needs domain-specific labels (for example, categorizing support tickets into internal categories unique to the business), the idea of “custom text classification” may appear. At AI-900 level, you are not required to configure training pipelines, but you should recognize that “custom” implies providing labeled examples and evaluating performance, while “prebuilt” implies using a ready model.

Also know the typical NLP “inputs and outputs” pattern: you send text, you receive structured results like confidence scores, detected entity types, offsets (where in the text the entity appears), and sometimes a document-level score (sentiment). This helps in exam scenarios that ask what the output looks like. If you see “returns JSON with entity types and confidence,” that’s a clue you’re in text analytics territory.

Exam Tip: In service-selection questions, highlight what is being asked: “extract insights” (analytics), “have a conversation” (bot), or “generate new content” (generative). The word “extract” is a strong signal for text analytics rather than generative AI.

Common trap: mixing OCR (computer vision) with text analytics. OCR is for turning images of text into characters. Text analytics is for understanding the meaning of already-available text. If the scenario starts with scanned PDFs or photos, you may need OCR first (Domain 3), then language analysis (Domain 4). AI-900 likes multi-step reasoning even if the question only asks for the best “next” service.

Section 5.3: Conversational AI overview: bots, orchestration, and user experience basics

Section 5.3: Conversational AI overview: bots, orchestration, and user experience basics

Conversational AI refers to systems that interact with users through natural language in chat or voice channels. On AI-900, you’ll see terms like bots and copilots. A bot is a conversational application that can answer questions, guide users through tasks, or hand off to a human agent. A copilot often implies a more assistive, context-aware experience embedded in a workflow (for example, drafting responses, summarizing, or helping complete tasks), commonly powered by generative AI behind the scenes.

The exam tests fundamentals of how conversations are managed. Typical components include: (1) a channel (web chat, Teams, SMS), (2) orchestration (routing user messages to the right logic), (3) understanding user input (detecting intent and entities), and (4) response generation (templated, retrieved, or generated). Even if you aren’t building the solution, you should be able to describe why intent/entity extraction matters: it turns unstructured utterances into structured inputs your business process can use.

User experience basics are also fair game. Good conversational design includes clear prompts (“What date would you like?”), confirmation (“I heard Seattle tomorrow—correct?”), and graceful fallback when confidence is low (“I’m not sure I understood; here are some options…”). These patterns reduce errors and support responsible AI by avoiding misleading outputs in high-stakes situations.

Exam Tip: If a question emphasizes “dialog,” “conversation flow,” “handoff to an agent,” or “multi-turn,” think conversational AI. If it emphasizes “extract entities from customer reviews,” think text analytics. The presence of multi-turn interaction is the giveaway.

Common trap: assuming every chatbot must be generative. Many bots are primarily retrieval- and rules-based (FAQ, knowledge base, guided forms). Generative responses may be added, but the exam often expects you to recognize that conversational AI can be built with structured dialog and intent/entity recognition—especially when accuracy and compliance matter.

Section 5.4: Generative AI fundamentals: LLMs, embeddings, retrieval concepts (high level)

Section 5.4: Generative AI fundamentals: LLMs, embeddings, retrieval concepts (high level)

Generative AI produces new content—text, summaries, explanations, or code—based on a prompt. The exam expects you to know what a foundation model is: a large model pretrained on broad data that can be adapted to many tasks. When the foundation model is specialized for language, it’s commonly called a large language model (LLM). Unlike classic NLP “extraction” tasks, generative AI outputs are probabilistic and may vary from run to run.

Two key terms: tokens and prompts. Tokens are chunks of text the model reads and writes (not always whole words). Token counts affect cost, latency, and maximum input/output size. A prompt is the instruction plus context you provide. Prompting basics are testable: being specific, providing context, and stating the desired format typically improves results.

The exam also touches retrieval concepts at a high level, especially the idea of grounding model outputs in your own data. This often appears as “retrieve relevant documents and then generate an answer,” which is the intuition behind retrieval-augmented generation (RAG). Related concept: embeddings—numeric vector representations of text that capture semantic similarity. Embeddings enable “find similar” search (semantic search), clustering, and retrieval of relevant passages to feed into a generative model.

Exam Tip: When a scenario says “answer questions using our internal policies” or “use company knowledge base,” the best conceptual approach is retrieval + generation (grounding), not “train a new LLM from scratch.” AI-900 often rewards the simplest correct architecture.

Common trap: confusing embeddings with prompts. Prompts are instructions/context sent to an LLM for generation. Embeddings are vectors used for similarity and retrieval; they are typically stored and searched (for example, against a vector index). Another trap is assuming generative AI is always correct—on the exam, you should acknowledge limitations like hallucinations and the need for grounding and safety controls.

Section 5.5: Azure OpenAI concepts: prompts, completions/chat, responsible use and safety

Section 5.5: Azure OpenAI concepts: prompts, completions/chat, responsible use and safety

Azure OpenAI Service provides access to powerful generative models hosted on Azure with enterprise considerations (security, regional deployment options, and integration with Azure services). For AI-900, focus on core interaction patterns: completions (the model continues text) and chat (messages with roles like system/user/assistant). Chat-based prompting is common because it supports multi-turn context and clearer instructions through role separation.

Prompt engineering patterns show up in scenario language even when not named explicitly. Examples include: setting a role or constraints (“You are a helpful support agent”), specifying output format (“Return a bulleted list”), providing examples (few-shot prompting), and adding grounding context (policy excerpts). The goal is reliability and usefulness, not model training.

Responsible AI and safety are heavily emphasized. You should know why guardrails matter: to reduce harmful content, protect privacy, and prevent unsafe instructions. At a fundamentals level, think in terms of: content filtering, moderation, limiting sensitive data in prompts, and using human review for high-impact decisions. If asked how to reduce hallucinations, strong answers include grounding with trusted sources, asking for citations/quotes from provided text, and constraining the model to answer only from supplied context.

Exam Tip: Watch for wording like “ensure the model does not generate hateful content” or “prevent leakage of confidential data.” These are governance/safety requirements. The correct answer is rarely “increase model temperature” or “use a bigger model”; it’s usually safety controls, data handling, and grounding.

Common trap: thinking temperature is a “safety” setting. Temperature primarily affects randomness/creativity, not compliance. Another trap is mixing up “fine-tuning” with “prompting.” Fine-tuning means adjusting a model with training data; prompting is just providing instructions/context at inference time. AI-900 tends to favor prompting and retrieval-based grounding as first-line approaches.

Section 5.6: Domains 4-5 exam-style questions and scenario decision practice

Section 5.6: Domains 4-5 exam-style questions and scenario decision practice

For Domains 4-5, many questions are “which service or approach should you use?” The fastest way to score is to classify the scenario into one of three buckets: (1) analyze/extract insights from text (sentiment, entities, key phrases, summarization), (2) converse with users across multiple turns (bot/coplanar experience with intent/entities and orchestration), or (3) generate new content (LLMs via Azure OpenAI, prompt patterns, grounding).

When you read a scenario, underline the verb: “detect,” “extract,” “classify,” “summarize,” “chat,” “draft,” “generate,” “rewrite,” “answer using our documents.” Then ask: is the expected output structured facts (analytics) or open-ended language (generative)? If it’s structured, prebuilt language analytics is often best. If it’s open-ended but must be accurate to company policy, the best decision pattern is retrieval + generation (grounding with internal content).

Exam Tip: Eliminate extreme options. “Train a custom model from scratch” is almost never the best first choice in AI-900 scenarios. Also eliminate solutions that ignore the input modality (for example, proposing language analytics when the input is an image that still needs OCR).

Safety and responsibility can be the deciding factor between two plausible answers. If the scenario involves customer-facing output, regulated industries, or sensitive data, prioritize solutions that support filtering, auditing, and controlled data handling. A common test angle is “what should you do to reduce risk?” Strong fundamentals answers include: do not include secrets in prompts, implement human-in-the-loop for high impact, use grounding and citations, and apply content moderation.

Finally, be ready for “match the task” items: sentiment aligns to opinion polarity; key phrase extraction aligns to topic summarization; entity recognition aligns to identifying named items; summarization aligns to shortening text; chat aligns to multi-turn assistance; prompts/tokens align to generative AI operation and constraints. If you can map each description to the right workload category, you’ll consistently pick the correct option even in unfamiliar business contexts.

Chapter milestones
  • Identify NLP tasks: sentiment, key phrases, entities, summarization
  • Understand conversational AI concepts (bots and copilots) at fundamentals level
  • Explain generative AI basics: foundation models, tokens, and prompts
  • Apply prompt engineering patterns and safety concepts
  • Exam-style practice set: NLP + generative AI on Azure
Chapter quiz

1. A retail company wants to automatically analyze thousands of product reviews to determine whether each review is positive, negative, mixed, or neutral. Which NLP task should you identify for this requirement?

Show answer
Correct answer: Sentiment analysis
Sentiment analysis classifies text by opinion polarity (positive/negative/mixed/neutral), which matches the requirement. Entity recognition extracts named items (for example, people, organizations, locations) but does not classify opinion. Key phrase extraction identifies important terms in the text, but it does not determine the reviewer’s sentiment.

2. A company wants to build an application that can summarize long customer support tickets into a short paragraph for agents to review. Which capability is being requested?

Show answer
Correct answer: Text summarization
Text summarization produces a shorter version of the original text while preserving meaning, which is exactly the scenario. Language detection identifies the language (for example, English vs. Spanish) and does not create summaries. Intent recognition is used in conversational AI to classify a user’s goal from an utterance, not to condense documents.

3. You are designing a basic customer service chatbot. A user types: "I need to reset my password." In conversational AI terminology, what is the most likely intent?

Show answer
Correct answer: ResetPassword
An intent represents the user’s goal (for example, ResetPassword), which best fits the utterance. "Password" is more like an entity/value extracted from text (a thing mentioned), not the goal. "User" is typically an actor/role or could be an entity in some designs, but it does not represent what the user is trying to accomplish.

4. A team is using a generative AI model to draft emails. They want to reduce the chance of the model returning harmful or policy-violating content. Which approach aligns with responsible generative AI fundamentals on Azure?

Show answer
Correct answer: Apply content safety filters/guardrails and use clear system instructions
Responsible use emphasizes safety guardrails, content filtering, and well-defined instructions to steer behavior and reduce harmful output. Increasing temperature generally increases randomness/creativity and can make outputs less predictable, not safer. Key phrase extraction is an NLP analysis task; it does not by itself prevent unsafe generated content.

5. You are preparing prompts for a foundation model. Which statement best describes what a token is in the context of generative AI?

Show answer
Correct answer: A unit of text (such as a word or part of a word) used by the model for input and output counting
Tokens are the pieces of text (often sub-words) that a model processes, and they are used to measure prompt/response length and limits. A pre-trained model used broadly is a foundation model, not a token. Mapping utterances to intents is part of conversational AI/NLU and is unrelated to the token concept.

Chapter 6: Full Mock Exam and Final Review

This chapter is your “dress rehearsal” for AI-900. Instead of teaching brand-new material, it helps you prove you can recognize which Azure AI capability matches a scenario, choose the best answer under time pressure, and avoid the exam’s most common wording traps. The AI-900 exam rewards clear categorization: is the scenario machine learning vs. prebuilt AI? Is it computer vision vs. language? Is it classic NLP vs. generative AI? And where do responsible AI principles fit into the decision?

You will work through two mixed-domain mock exam passes (without seeing any actual questions in this text), then complete a weak-spot analysis and a final domain-by-domain rapid recap strategy. Throughout, you’ll practice the most important test skill: selecting the “best” answer, not merely a “true” answer, by anchoring on requirements (modality, output type, latency, and governance needs) and the correct Azure service family.

Exam Tip: AI-900 questions often include extra details. Treat each scenario like a requirements list: identify the input type (text, image, audio), the task (classify, extract, generate, predict), and the constraint (must explain predictions, must run in real time, must be responsible/secure). Then map to the service.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final review: domain-by-domain rapid recap: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final review: domain-by-domain rapid recap: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam rules, pacing plan, and how to review mistakes

Section 6.1: Mock exam rules, pacing plan, and how to review mistakes

Run your mock exam like the real test: single sitting, no notes, no pausing for “just a quick search.” The goal is to build decision speed and reduce second-guessing. Use a pacing plan: first pass to answer everything you can confidently, second pass for medium-confidence items, final pass for flagged questions only. This prevents you from burning time early and rushing later, which is a top reason capable candidates underperform.

During the mock, practice “service family first” thinking. When a scenario describes extracting key phrases, sentiment, entities, or PII detection, mentally label it as Language (text analytics). If it describes object detection, OCR, or image classification, label it as Vision. If it asks about training vs inference, features/labels, evaluation metrics, or pipelines, label it as machine learning. If it asks about generating text, summarizing, or chat-style responses from a foundation model, label it as generative AI and Azure OpenAI concepts.

Reviewing mistakes is where improvement happens. For each missed item, write a one-line reason in one of three categories: (1) concept gap (didn’t know), (2) misread requirement (rushed), or (3) best-answer trap (two plausible options). Your re-study plan should focus more on (2) and (3) than you expect—AI-900 is often a reading and mapping exam.

Exam Tip: When reviewing, don’t just note the correct service; note the “disqualifier” for the wrong choice (e.g., “needs OCR → speech service can’t help,” or “needs custom training → prebuilt model isn’t the best fit”). That disqualifier is what you’ll recognize under pressure.

Section 6.2: Mock Exam Part 1 (mixed domains, exam-style)

Section 6.2: Mock Exam Part 1 (mixed domains, exam-style)

Mock Exam Part 1 should intentionally mix domains so you practice quick switching. Expect items that test whether you can separate “AI workload type” from “Azure product name.” For example, you may see business scenarios (customer support, document processing, quality inspection, demand forecasting) where the exam wants you to choose between prebuilt AI services and custom ML. Your job is to identify what must be learned from data versus what can be handled by a ready-made model.

Key exam objectives this part typically stresses: training vs inference, features vs labels, and model evaluation. If a scenario describes building a predictive model from historical data (e.g., predicting churn), that implies supervised learning with labeled outcomes. If it describes running a model to produce predictions on new data, that’s inference. Watch for questions that use “deploy,” “consume,” “score,” or “real-time endpoint”—those words usually point to inference rather than training.

You should also be ready to map classic NLP tasks to Azure Language capabilities. The exam likes tasks such as sentiment analysis, key phrase extraction, entity recognition, language detection, and summarization. In exam-style wording, the task is often described as an outcome (“identify positive/negative feedback,” “detect names and locations,” “remove personal data”). Focus on the output: sentiment score, entities, key phrases, or redacted text.

Exam Tip: If the scenario involves extracting text from images or PDFs, mentally split it into two steps: (1) OCR/read text (Vision/Document Intelligence), then (2) analyze the extracted text (Language). Many candidates pick only the text analytics tool and forget the OCR prerequisite.

Section 6.3: Mock Exam Part 2 (mixed domains, exam-style)

Section 6.3: Mock Exam Part 2 (mixed domains, exam-style)

Mock Exam Part 2 should continue mixing domains but increase emphasis on computer vision, conversational AI concepts, and generative AI. For vision, be clear on the difference between image classification (what is in the image), object detection (where objects are, typically bounding boxes), and OCR (extract text). The exam often tests whether you can choose the correct workload based on wording like “count items on a shelf” (detection) versus “identify if a photo contains a cat” (classification) versus “read a serial number” (OCR).

For conversational AI, AI-900 often stays conceptual: intent recognition, utterances, entities/slots, and the overall goal of turning natural language into an actionable decision. If a prompt describes “users ask questions and the system routes requests,” think intents and entities. If it describes “answers from a knowledge base,” look for Q&A/knowledge retrieval phrasing. If it describes “multi-turn conversation,” focus on context and dialog flow rather than just single-message text analytics.

Generative AI questions frequently check that you know what foundation models are, what prompt engineering is (instructions, context, examples, constraints), and how Azure OpenAI fits as a managed offering. You may need to differentiate between using a pre-trained generative model (no training required for basic use) versus fine-tuning/customization, and between retrieval-augmented generation (bringing your own data via search) and “the model already knows it.”

Exam Tip: When you see a generative AI scenario, scan for governance cues: safety filters, content moderation, data privacy boundaries, and responsible AI. If two answers both “work,” the better answer often includes controls for safety, monitoring, or transparency.

Section 6.4: Scoring your results and creating a targeted re-study plan

Section 6.4: Scoring your results and creating a targeted re-study plan

After both mock parts, score yourself by domain rather than by total alone. Create a simple grid with five rows aligned to the course outcomes: AI workloads and responsible AI; ML fundamentals; computer vision; NLP; generative AI. Mark each missed item with the reason type: concept gap, misread requirement, or best-answer trap. This turns a disappointing score into a practical map of what to fix in the next 48–72 hours.

Next, build a targeted re-study plan that prioritizes “high yield, easy wins.” If you missed items because you confused detection vs classification, that’s quick to correct and likely to reappear. If you missed training vs inference terminology, that’s foundational and must be fixed first. If you missed because you ignored a word like “extract text” or “bounding box,” your plan should include deliberate reading drills: practice underlining the single phrase that defines the required output.

Re-study should be active. Don’t reread notes passively; instead, write mini “if-then” rules: “If the output is redacted PII, then choose a language feature for PII detection.” “If the requirement is location of objects, then detection, not classification.” For responsible AI, list the principles you’re expected to recognize: fairness, reliability & safety, privacy & security, inclusiveness, transparency, and accountability.

Exam Tip: Track your “flip-flops.” If you changed an answer and got it wrong, your issue is confidence calibration. Your fix is a rule: only change when you can cite a requirement you missed the first time.

Section 6.5: High-frequency traps and best answer selection techniques

Section 6.5: High-frequency traps and best answer selection techniques

AI-900 is full of “nearly right” options. The exam tests whether you can choose the most appropriate service or concept, not merely something related to AI. One frequent trap is confusing prebuilt cognitive services with custom machine learning. If the scenario needs a standard capability (OCR, sentiment, basic entity extraction), prebuilt services are usually best. If it needs a prediction unique to your business (custom risk score from your internal features), that points to training a model with Azure Machine Learning concepts.

Another trap is mixing modalities. Text analytics does not read pixels; vision does not infer sentiment from a paragraph unless the text is extracted first. Similarly, “speech” is its own modality: transcribing audio is not the same as analyzing text sentiment, even if the end product is words.

Best-answer technique: eliminate options by checking four filters. (1) Input type: text vs image vs audio. (2) Output type: classify, detect, extract, generate, predict. (3) Build vs buy: custom training needed or not. (4) Governance: responsible AI, privacy, explainability, monitoring. This approach is fast and consistent.

  • Trap: “Accuracy” vs “precision/recall.” If the scenario emphasizes missed detections (false negatives), recall matters; if it emphasizes avoiding false alarms (false positives), precision matters.
  • Trap: “Training dataset” vs “test dataset.” Training is for learning; testing/validation is for evaluation. If the prompt says “evaluate,” do not pick training steps.
  • Trap: “Prompting” vs “fine-tuning.” Prompt engineering is shaping instructions and context; fine-tuning changes model behavior via additional training.

Exam Tip: When two answers seem plausible, re-read the question stem and locate the single verb that matters most: “detect,” “extract,” “generate,” “predict,” or “translate.” The verb usually reveals the correct workload category.

Section 6.6: Exam day checklist: environment, time management, and confidence routine

Section 6.6: Exam day checklist: environment, time management, and confidence routine

On exam day, remove preventable stress. Confirm your testing environment (quiet room, stable internet if remote, allowed ID, and a clean desk). If you’re taking the exam online, do a system check early and close unnecessary apps. If you’re at a test center, arrive early enough to settle in without rushing—rushing increases misreads, which are the most expensive errors on AI-900.

Use a simple time management routine: start with a calm first pass to secure easy points, flag anything that requires heavy reading, and keep moving. Don’t attempt to “perfect” the first five questions—momentum matters. If you feel stuck, force an elimination step: remove any option that mismatches modality (text/image/audio) or required output (extract vs detect vs generate). Then choose the best remaining answer and move on.

Confidence routine: before you begin, remind yourself of the core map you’ve built across the course outcomes—AI workloads, ML fundamentals (training/inference, features/labels, evaluation), vision tasks (classification/detection/OCR), NLP tasks (sentiment/entities/key phrases/conversation concepts), and generative AI (foundation models, prompt basics, Azure OpenAI). This mental map prevents panic because every question must land in one of these buckets.

Exam Tip: If you finish early, use remaining time to re-check only flagged questions. Re-reading everything can cause unnecessary answer changes. Your goal is to catch misreads, not to talk yourself out of correct choices.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
  • Final review: domain-by-domain rapid recap
Chapter quiz

1. A customer support team wants an Azure AI solution that can answer questions using the company’s policy documents and past tickets. The answers must be generated in natural language and should be grounded in the provided content to reduce hallucinations. Which approach best fits the requirement?

Show answer
Correct answer: Use Azure OpenAI with Retrieval Augmented Generation (RAG) over the company’s indexed documents
Azure OpenAI with RAG is designed for generative Q&A grounded on your enterprise content (classic AI-900 mapping: generative AI + retrieval). Azure AI Vision is for image analysis (OCR/object detection) and does not provide grounded conversational answers by itself. Azure Machine Learning regression predicts numeric values (like time) and is not the best fit for generating natural-language answers from documents.

2. You need to categorize incoming emails into predefined topics (Billing, Technical Support, Sales). You want to use a prebuilt service with minimal training and no code-heavy model development. Which Azure AI capability should you choose?

Show answer
Correct answer: Azure AI Language (text classification)
Topic labeling of text is a Natural Language Processing classification task, which fits Azure AI Language. Azure AI Vision is for analyzing images, not email text. Azure OpenAI can generate text, but using a generative model is not the most direct prebuilt choice for deterministic topic classification in the AI-900 service mapping.

3. A company wants to extract printed invoice numbers and totals from scanned PDF invoices. The solution must return the recognized text so it can be stored in a database. Which Azure AI service is the best match?

Show answer
Correct answer: Azure AI Vision (OCR / Read)
Extracting text from scanned documents is an OCR use case, which maps to Azure AI Vision (Read/OCR). Key phrase extraction requires existing digital text and does not perform text recognition from images. Anomaly detection in Azure Machine Learning is for identifying unusual patterns in numeric/time-series data, not reading invoice text.

4. You are reviewing a proposed AI solution for loan approvals. Leadership asks which responsible AI principle is most directly addressed by providing a reason code for each approval/denial decision. What principle is this primarily supporting?

Show answer
Correct answer: Transparency and explainability
Providing reasons for decisions maps to transparency/interpretability (explainability), a core responsible AI topic in AI-900. Reliability and safety focuses on consistent performance and avoiding harmful failures, not explaining individual predictions. Privacy and security focuses on protecting data and access controls rather than explaining model outputs.

5. During an AI-900-style scenario question, you are given extra details (company size, office locations, and device brands). The real requirement is: process live audio from a call center and convert it to text in near real time. Which Azure AI capability should you select?

Show answer
Correct answer: Azure AI Speech (speech-to-text)
Live audio transcription is a speech modality requirement; Azure AI Speech provides real-time speech-to-text. Image captioning is for describing images, not audio. Sentiment analysis operates on text (often after transcription) and is not the primary service to convert audio into text.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.