AI Certification Exam Prep — Beginner
Timed AI-900 sims plus targeted drills to fix weak areas fast.
This course is a focused, simulation-first blueprint for learners preparing for Microsoft’s AI-900: Azure AI Fundamentals exam. Instead of long lectures, you’ll use timed mock exams and targeted weak-spot repair to build confidence with the official objectives and the way Microsoft asks questions. The result: faster recall, better service selection, and fewer avoidable mistakes under time pressure.
You’ll work through a 6-chapter “book” structure. Chapter 1 sets you up with the exam mechanics (registration, delivery options, scoring expectations, and pacing strategy). Chapters 2–5 align directly to the official AI-900 exam domains and include exam-style drills designed to mirror real objective wording. Chapter 6 delivers a full mock exam experience and a structured final review loop.
If you’re new to certifications, you’ll also learn how to read Microsoft-style prompts, eliminate distractors, and avoid common traps (like confusing similar service families or picking an overly complex solution).
This blueprint maps to the published AI-900 domains and keeps practice anchored to objective names:
Each domain chapter includes deep explanation at the level expected for AI-900 and then pivots quickly into exam-style practice. After each drill set, you’ll do a short review loop: (1) identify which objective the question tested, (2) record why your chosen answer was tempting, and (3) write a one-sentence rule you can reuse on the next similar prompt.
Chapter 6 brings everything together with two timed simulations and a structured weak-spot analysis process. You’ll end with a last-pass checklist that covers pacing, reviewing flagged questions, and keeping your choices aligned to “fundamentals” scope (a common pitfall is selecting services or techniques beyond what the exam expects).
This course is designed for beginners with basic IT literacy and no prior certification experience. You don’t need to be a developer or data scientist. You’ll learn to recognize scenarios and pick the most appropriate Azure AI approach—exactly what AI-900 tests.
Create your learning plan and start practicing today. Register free to track progress and revisit weak domains, or browse all courses to compare certification paths.
By the end, you’ll have completed domain-aligned drills, two full timed simulations, and a repeatable weak-spot repair routine—so you can walk into the Microsoft AI-900 exam with realistic timing practice and clear decision rules for every domain.
Microsoft Certified Trainer (MCT)
Jordan McAllister is a Microsoft Certified Trainer who builds beginner-friendly certification prep for Azure fundamentals tracks. He specializes in translating Microsoft exam objectives into focused practice and repeatable test-taking strategies for first-time candidates.
AI-900 is a fundamentals exam, but “fundamentals” in Microsoft exams still means you must make consistent, test-aligned decisions: identify the workload, choose the right Azure service category, and recognize responsible AI considerations. This course is built as a Mock Exam Marathon, so your job in Chapter 1 is to set up the rules of the game: understand the format and policies, build a schedule that prevents cram-burnout, and establish a baseline so practice time targets your weak spots instead of repeating what you already know.
Across this book, you’ll map common scenarios to the official domains: describing AI workloads (including responsible AI), machine learning principles on Azure, computer vision workloads, NLP workloads, and generative AI workloads with Azure OpenAI patterns. The exam rewards clarity: if you can name the task (classification, OCR, entity extraction, chat completion) and match it to the correct service family, you will score well. The rest is execution: timing, reading precision, and a disciplined review loop.
Exam Tip: Treat AI-900 like a vocabulary-and-decision exam. When you miss a question, don’t just note the right option—write the rule that would have made the choice obvious (e.g., “OCR → Azure AI Vision Read,” “custom model training → Azure Machine Learning,” “chatbot orchestration → Azure AI Bot Service + Language/Orchestrator,” “generative text → Azure OpenAI”).
This chapter walks you through exam orientation and a study game plan aligned to official objectives, then sets up your practice environment and baseline assessment so your timed sims are diagnostic, not just repetitive drilling.
Practice note for Understand AI-900 format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set a realistic schedule: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a study plan aligned to official domains: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Set up your practice environment and baseline assessment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand AI-900 format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set a realistic schedule: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a study plan aligned to official domains: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Set up your practice environment and baseline assessment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 measures whether you can recognize common AI scenarios and select the appropriate Azure AI approach at a high level. It is not a coding exam, but it is also not purely definitional. You will repeatedly be asked to interpret a short business scenario, identify the AI workload type, and pick the best Azure service category or concept. The core domains you should map your study plan to are: (1) describing AI workloads and responsible AI, (2) fundamental ML principles on Azure (types of ML, training vs inference, features/labels, evaluation), (3) computer vision workloads (image analysis, object detection concepts, OCR), (4) NLP workloads (sentiment, key phrases, entity recognition, translation, conversation), and (5) generative AI workloads (LLMs, prompts, completions, embeddings, responsible use).
What the exam “tests for” is judgment: can you choose the most direct capability for the task? For example, if the requirement is to extract text from an image, your mental path should be “computer vision → OCR → Azure AI Vision.” If the requirement is to build and train a custom model with experiments, compute, datasets, and pipelines, think “Azure Machine Learning.” If the requirement is to analyze user reviews for sentiment or extract entities, think “Azure AI Language.” For generative scenarios, you should recognize when the task is content generation or summarization (Azure OpenAI), versus when it is classic NLP classification/extraction (Azure AI Language).
Exam Tip: Practice translating the scenario into a single verb-noun task statement before looking at options: “classify images,” “extract printed text,” “summarize emails,” “identify PII,” “predict house prices.” That one line often reveals the correct service family.
A frequent trap is confusing “AI workloads” with “Azure products.” The exam expects you to first name the workload (vision, language, ML, gen AI) and only then select the best-fitting service. If your approach starts with scanning for product names, you’ll be pulled into distractors that sound plausible but don’t match the requirement.
Your score improves when logistics stop being a variable. Register early, pick a date that supports spaced repetition, and choose a delivery option (test center vs online proctored) that matches your environment. Online proctoring is convenient but unforgiving: unstable internet, dual-monitor setups, background noise, and unapproved materials can end a session. Test center delivery reduces home-environment risk but adds travel and scheduling constraints. Either way, plan your study schedule backward from the exam date with milestones: baseline diagnostic, domain study blocks, timed sims, and final weak-spot repair.
Policies matter because they affect your performance psychology. You should know what identification is required, when you can reschedule, and how check-in works. If you choose online proctoring, do a device and network check well before exam day, and prepare a clean workspace. If you choose a test center, confirm location, arrival time, and what you can bring. The goal is to prevent “surprise stress” that steals time and focus.
Exam Tip: Book your exam for a time of day when you reliably do focused reading. AI-900 questions are short, but the trick is subtle wording; you want your best attention, not your last energy.
Build a realistic schedule: for many learners, 2–4 weeks with consistent practice is better than a single intense weekend. Allocate time across all domains even if you have prior experience—fundamentals exams often include breadth questions that punish gaps (for example, responsible AI principles or the difference between classic NLP and generative AI patterns).
Microsoft exams typically report scores on a scaled range, and the “cut score” is the minimum passing threshold. You don’t need to chase perfection; you need dependable accuracy across domains. The practical implication is that your study plan should aim for balanced competence, not peak performance in one area and weakness in another. In fundamentals exams, a few missed questions in a small domain can be the difference between pass and fail if you ignored that topic entirely.
Because you won’t see “partial credit” in the way you study, train yourself to make clean decisions. When reviewing practice results, categorize misses into three buckets: (1) concept gap (you didn’t know the definition or service), (2) misread (you missed a keyword like “real-time,” “custom,” “translated,” “handwritten”), and (3) overthinking (you changed from the correct basic choice to an advanced-sounding distractor). Your retake strategy—ideally never needed—depends on which bucket dominates.
Exam Tip: Track your “avoidable errors” rate (misreads + overthinking). If that rate is high, more content study won’t fix it; you need timed reading drills and a stricter answer-selection process.
If you do need a retake, treat it as a targeted repair cycle: re-run a baseline diagnostic, focus on weak domains, and do timed simulations that mirror the real constraint. Avoid taking the exam again with the same approach; change the system (study loop, timing strategy, error log) rather than simply adding hours.
This course is a Mock Exam Marathon: your main training tool is timed simulations (“timed sims”) followed by disciplined review loops. The timed sim builds the two things content reading cannot: pacing and decision-making under pressure. Start with a baseline timed sim early—even if you feel unprepared—because it reveals which domains and question styles cost you time. Then use short, frequent sims (or timed sets) to practice reading precision and to stabilize your approach.
Your review loop is where score gains happen. After each sim, do not merely check correct answers. Instead, write an error log entry with: the workload type, the key requirement word(s), the correct service/concept, and the reason your choice was wrong. Over time, you’ll see patterns: confusing OCR with image analysis, mixing Azure AI Language with Azure OpenAI, or forgetting responsible AI principles. Convert those patterns into “one-line rules” you can recall under time pressure.
Exam Tip: When a question feels unfamiliar, anchor to the task type and the simplest matching service. Fundamentals exams reward direct mapping more than architectural creativity.
Use your practice environment to reinforce concepts: explore Azure AI service names and what they do at a high level, but don’t fall into the trap of trying to “learn everything in the portal.” Your goal is scenario-to-solution mapping, not implementation mastery.
Microsoft objective-style questions are engineered with plausible distractors. The most common trap is “almost right, but not for this requirement.” Train yourself to spot requirement keywords and eliminate options that violate them. Examples of requirement keywords include: “custom model,” “no code,” “real-time,” “extract text,” “detect faces,” “translate,” “summarize,” “responsible,” “explainable,” and “private data.” If an option is powerful but doesn’t directly satisfy the requirement, it’s a distractor.
Another trap is mixing service families that sound similar. “Azure Machine Learning” (training and managing ML) is not the same as using “Azure AI services” (prebuilt vision/language/speech capabilities). “Azure AI Language” (classical NLP tasks) is not the same as “Azure OpenAI” (LLM-driven generation and chat). The exam often tests whether you can keep these boundaries clear.
Exam Tip: Use a two-pass elimination: (1) remove anything that doesn’t match the workload (vision vs language vs ML vs gen AI), then (2) remove anything that doesn’t match the constraint (prebuilt vs custom, batch vs real-time, extract vs classify).
Watch for “scope creep” traps where a complex-sounding solution is offered for a simple task. Fundamentals questions usually want the most straightforward service that meets the need, not a multi-component architecture. Also beware of negative wording (“NOT,” “least likely,” “which is false”). Slow down on those and rephrase them into a positive statement before answering.
Finally, responsible AI content can appear as a scenario “add-on”: fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. Distractors often swap these terms; pick the principle that matches the harm described (e.g., demographic bias → fairness; inability to understand model decisions → transparency; misuse of personal data → privacy and security).
Your first actionable deliverable is a baseline diagnostic that produces a personal weak-spot map. Take a timed simulation under realistic conditions and capture two outputs: (1) domain-level performance (which objective areas are lowest) and (2) error-type performance (concept gaps vs misreads vs overthinking). This is how you “set up your practice environment and baseline assessment” so future study time is targeted.
Build your weak-spot map as a simple table or checklist with these columns: objective area, subtopic, confidence (high/medium/low), common trap, and next action. For example: “Computer vision → OCR: low confidence; trap: confusing Read vs general image analysis; action: review OCR use cases + do a timed set focused on text extraction scenarios.” Do the same for ML fundamentals (supervised vs unsupervised vs reinforcement; classification vs regression; training vs inference) and for generative AI (prompting basics, embeddings concept, responsible use patterns).
Exam Tip: Don’t label a domain “strong” just because you got questions right. If you guessed, mark it as weak. The exam punishes fragile knowledge because wording variations break guessing strategies.
Your study schedule should now be realistic and data-driven: allocate the majority of time to the bottom two domains while maintaining light rotation through the others to prevent decay. End each week with a timed sim to validate that weak-spot repairs are transferring into performance. If your score stagnates, check whether the issue is content or execution (timing and reading). This weak-spot map becomes the backbone of the entire marathon.
1. You are planning your AI-900 preparation. Your goal is to maximize score improvement with limited study time. Which approach BEST aligns with the AI-900 exam’s emphasis and this course’s “Mock Exam Marathon” strategy?
2. A company asks you to set expectations for AI-900. They want to know the best way to avoid losing points due to misreading questions. What should you recommend?
3. A startup wants to build a study plan aligned with AI-900’s official objectives. Which set of study domains MOST closely matches how the exam content is organized?
4. You are reviewing a missed practice question: “Extract printed and handwritten text from scanned invoices.” Which rule should you record to make the correct choice obvious on future AI-900 items?
5. Your manager wants you to set up a practice environment and schedule for AI-900. You must reduce cram-burnout risk and ensure practice results are diagnostic. What should you do FIRST?
Domain 1 of AI-900 rewards a specific skill: reading a scenario and quickly classifying the workload, the AI approach, and the Azure service family that fits—without over-engineering. Most missed points happen when candidates confuse “what the business wants” (scenario) with “how it’s implemented” (service) or when they treat every problem as machine learning. This chapter builds a repeatable mapping method: identify the workload category, decide whether the solution is predictive, generative, or analytical, then select a high-level Azure AI service family. Along the way, you’ll layer in Responsible AI requirements because many questions hide the real ask behind compliance, privacy, or transparency cues.
This chapter also supports your mock-exam marathon approach: you’ll practice fast triage (“what is this asking?”), avoid common traps (“custom model” vs “prebuilt API”), and use weak-spot analytics by tagging mistakes to one of the workload categories. Exam Tip: When you miss a question, don’t just note the right answer—label the mistake as (1) workload misclassification, (2) approach confusion, (3) service-family mismatch, or (4) Responsible AI oversight. Those four labels predict your score gains better than re-reading notes.
Practice note for Recognize AI workloads and when AI is appropriate: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match workload types to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles in exam scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed domain drill set: mixed single/multiple-answer questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize AI workloads and when AI is appropriate: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match workload types to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles in exam scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed domain drill set: mixed single/multiple-answer questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize AI workloads and when AI is appropriate: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match workload types to Azure AI services at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
The exam expects you to recognize AI workloads by the data type and the output style. Start by asking: “What is the input?” Images/video → vision. Text documents/chat messages → language (NLP). Audio streams → speech. Tabular events and rules-driven outcomes → decision. Prompts that request new content (text, code, images) → generative AI.
Vision workloads include image classification, object detection, image description, and OCR (extracting text from images). On AI-900, you’re rarely asked to design a CNN; you’re asked to name the workload correctly and choose the right service family. Language workloads include sentiment analysis, key phrase extraction, entity recognition, summarization, translation, and conversational understanding. Speech includes speech-to-text, text-to-speech, and translation of spoken language. DecisionGen AI
Common trap: candidates label any “prediction” as machine learning, but some scenarios are primarily information extraction. Example: “Extract invoice totals from scanned PDFs” is not forecasting; it’s OCR + document understanding. Another trap is confusing “language” with “gen AI.” If the output is a label or extracted fields (sentiment, entities), that’s analytical NLP. If the output is a newly written paragraph, a plan, or code, that’s generative.
Exam Tip: If the scenario mentions bounding boxes, tags, captions, OCR, or “read text in an image,” anchor on Vision. If it mentions intent, entities, sentiment, or summarization of existing text, anchor on Language. If it mentions “draft,” “compose,” “rewrite,” “chat,” or “generate,” anchor on Gen AI.
Domain 1 also checks whether you can separate three solution styles that sound similar in business language. Predictive solutions estimate an unknown value or class based on patterns in historical data (classification, regression, forecasting). Analytical solutions extract structure from unstructured content (labels, entities, key phrases, topics) or find patterns (clustering, anomaly detection) without necessarily “creating” content. Generative solutions produce new content conditioned on prompts and context (text, code, images) and often rely on large language models.
Look for cues. Predictive questions often reference training data, features, labels, or “predict next month’s sales,” “estimate churn,” “detect fraud,” “approve a loan.” Analytical questions use verbs like “extract,” “identify,” “classify text,” “recognize entities,” “summarize a document,” “detect language.” Generative questions use verbs like “write,” “draft,” “create,” “respond conversationally,” “generate a product description,” “create code,” “produce an image.”
Common trap: “summarize” can be analytical NLP (extractive or abstractive). On AI-900, summarization is typically mapped to Language services (summarization) unless the scenario explicitly frames it as a chat-based content-generation app with prompting and grounded context. Another trap: “recommendations” can be predictive/decision, not generative—recommendation engines rank items; they aren’t asked to write marketing copy.
Exam Tip: When two answers seem plausible, choose the one that matches the output type. If the output is a probability/score/label → predictive. If the output is extracted structure from existing content → analytical. If the output is novel free-form content → generative.
Many scenarios are really about how AI is used operationally: to assist a human, to automate a decision, or to route uncertain cases for review. The exam will often hide this behind phrases like “help agents triage,” “reduce manual work,” “flag for review,” or “automatically approve.” Your job is to infer the pattern and its implications for risk and Responsible AI.
Decision support means AI provides a recommendation or insight while a person remains accountable (for example, a call-center agent sees sentiment and suggested responses). Automation means the AI output directly triggers an action (approve/deny, block/unblock, route shipments). Human-in-the-loop (HITL) sits between them: the system automates low-risk/high-confidence cases and escalates low-confidence or high-impact cases to humans.
Why this matters on AI-900: “fully automated” answers are often wrong when the scenario has high stakes (healthcare, finance, employment) or when it explicitly requests explainability, auditing, or review workflows. HITL also aligns with reliability and accountability: you can reduce harm by setting thresholds, monitoring drift, and requiring manual review for edge cases.
Common trap: selecting a service based only on the data type while ignoring operational constraints. Example: a content moderation scenario is not just “language analysis”—it’s a safety workflow requiring policy thresholds and escalation. Another trap: assuming the model is always right; the exam frequently rewards acknowledging uncertainty (confidence scores) and review processes.
Exam Tip: If you see “assist,” “recommend,” “provide insights,” or “support,” think decision support. If you see “automatically,” “without human intervention,” or “real-time blocking,” check for Responsible AI cues and consider HITL unless the scenario is explicitly low risk.
Responsible AI is tested as principles applied to scenarios, not as definitions to memorize. Expect questions that ask what to do (or what principle is being addressed) when stakeholders worry about bias, outages, sensitive data, or “black box” decisions.
Fairness focuses on avoiding discrimination across groups (for example, a loan model that disadvantages a protected demographic). Typical mitigations include representative data, bias evaluation, and monitoring. Reliability & safety concerns consistent performance under expected conditions and safe behavior under unexpected inputs; think monitoring, fallbacks, and testing across edge cases. Privacy & security concerns handling personal/sensitive data appropriately (data minimization, access control, encryption, retention policies). Transparency is about making users aware they are interacting with AI, explaining limitations, and providing understandable reasons for outputs when required. Accountability means humans and organizations remain responsible: governance, audit trails, and clear ownership of outcomes.
Common trap: confusing transparency with explainability. Transparency includes disclosures (“this is AI,” “these are limitations”) and clarity about data use. Explainability is often part of transparency but is specifically about reasons behind a prediction. Another trap: treating privacy as optional when the scenario includes names, addresses, health records, or voice recordings—privacy becomes the primary constraint and may change the recommended architecture (for example, limiting data sent to external systems, enforcing retention).
Exam Tip: In scenario questions, underline the “worry statement.” If the stakeholder says “biased,” map to fairness. If they say “can’t go down,” map to reliability. If they say “PII” or “regulatory,” map to privacy. If they say “users must know,” map to transparency. If they say “who is responsible,” map to accountability and governance.
AI-900 stays high level: you’re mapping workloads to service families, not designing pipelines. Use a “family first” approach before you worry about any feature name. If the input is images/video → Azure AI Vision. If the input is text → Azure AI Language. If the input is audio → Azure AI Speech. If the user needs to find relevant content across documents → Azure AI Search. If the user wants generative chat/completions/embeddings → Azure OpenAI Service.
Azure AI Vision covers image analysis, detection, and OCR-style tasks. In exam scenarios, words like “read text from a photo,” “analyze a product image,” or “detect objects” point here. Azure AI Language supports sentiment, entities, classification, summarization, and question answering patterns over text. Azure AI Speech handles speech-to-text, text-to-speech, and translation for voice experiences. Azure AI Search supports indexing and retrieval over your data; it becomes especially important when the scenario describes “search across manuals,” “find policy clauses,” or “retrieve relevant passages.” Azure OpenAI Service is used for generative patterns like chat, completions, and embeddings-based retrieval augmentation.
Common trap: choosing Azure OpenAI for any text scenario. If the scenario just needs extraction (entities, sentiment) with deterministic outputs, Language is usually the best fit. Another trap: confusing Search with Language. Search retrieves; Language extracts and classifies. In modern architectures they combine (retrieve with Search, then generate with OpenAI), but the exam often asks for the primary capability the scenario highlights.
Exam Tip: If the question is “Which service analyzes?” think Vision/Language/Speech. If the question is “Which service finds relevant documents?” think Search. If the question is “Which service generates a response or content?” think OpenAI.
This course uses timed simulations, so your performance depends on speed and pattern recognition. Domain 1 questions are typically short, and the distractors are “near misses” (right family, wrong workload; right workload, wrong approach). Your drill strategy: (1) classify workload category, (2) classify solution style (predictive/analytical/generative), (3) identify the highest-level Azure family, (4) scan for Responsible AI constraints that override a default choice.
Rationale patterns you should practice (without memorizing specific questions):
Timing guidance for the marathon: do not over-read. For each item, spend the first 5–10 seconds highlighting nouns (data type) and verbs (task). Then spend 5 seconds eliminating options that don’t match the input modality (for example, Speech options for an image-only scenario). Finally, if two remain, use the output-type rule: label/score/extract vs generate. Exam Tip: If you can’t justify an option in one sentence (“This is OCR, so Vision”), it’s probably a distractor. Your justification sentence is your best defense against tricky wording.
Weak-spot repair: after each timed drill, bucket misses into (A) modality error (vision/language/speech), (B) approach error (predictive/analytical/generative), (C) service-family confusion (Search vs Language vs OpenAI), or (D) Responsible AI oversight. Then re-run only the bucket you missed most until your reaction becomes automatic—this is the fastest path to raising your Domain 1 score.
1. A retail company wants to automatically categorize customer emails into "billing issue," "return request," or "product question" so that tickets route to the correct team. The company has thousands of labeled examples from the past year. Which AI workload is most appropriate?
2. A healthcare provider wants to extract key fields (patient name, date of birth, and policy number) from scanned insurance forms that may vary slightly in layout. They want a prebuilt capability rather than training a custom model. Which Azure AI service family best fits?
3. A manufacturer collects temperature readings from equipment sensors and wants to detect unusual spikes that may indicate impending failure. The goal is to flag abnormal behavior rather than predict a specific numeric value. Which workload type is this?
4. A company wants to add a chat experience to its website that answers questions using the company’s policy documents and returns grounded responses with citations. Which Azure AI service family is the best high-level match?
5. A bank deploys an AI model that recommends whether to approve a loan. Auditors require the bank to provide understandable reasons for decisions and to detect potential bias across demographic groups. Which Responsible AI principle is most directly addressed by these requirements?
Domain 2 on AI-900 tests whether you can reason about machine learning (ML) at a practical, “choose-the-right-approach” level—without needing to code. Expect questions that describe a business scenario and ask you to identify the right ML type (classification vs. regression vs. clustering), the right stage (training vs. inference), or the right Azure Machine Learning capability (AutoML, pipelines, compute). This chapter is your weak-spot repair kit: we’ll anchor the vocabulary (features/labels, algorithms/models, training/inference), then build scenario recognition patterns, then lock in evaluation metrics and overfitting awareness at exam depth.
A consistent exam theme: the test often hides the answer in plain language. Words like “predict a category,” “estimate a numeric value,” “group similar items,” “detect unusual behavior,” or “run repeatedly at scale” map directly to ML types, metrics, and Azure services. Your job is to translate the story into ML fundamentals and choose the simplest correct option.
Exam Tip: If the scenario output is a number (price, demand, temperature), think regression. If the output is a label (fraud/not fraud, churn/not churn), think classification. If there is no labeled target and the task is “discover groups,” think clustering. If the task is “find unusual,” think anomaly detection.
Practice note for Master ML fundamentals: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose the right ML type for a given scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand evaluation metrics and overfitting at exam depth: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed simulation: ML-focused caselets and MCQs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Master ML fundamentals: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose the right ML type for a given scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand evaluation metrics and overfitting at exam depth: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed simulation: ML-focused caselets and MCQs: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Master ML fundamentals: features, labels, training, and inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose the right ML type for a given scenario: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to use ML terms precisely. A dataset is the collection of examples you learn from; each example is described by features (inputs) and may include a label (the correct output). The algorithm is the learning procedure (for example, logistic regression or decision trees). The model is the artifact produced after training—what you deploy and use to make predictions.
Training is the process of fitting a model to data (learning patterns). Inference is using that trained model to generate predictions on new data. Exam items frequently try to blur these two by describing “the system predicts…” while also mentioning “the model learns from…” in the same paragraph—your task is to identify what phase the system is in.
Another core term is feature engineering: selecting/transforming inputs so the model can learn effectively. On AI-900, you won’t be asked to craft code, but you may be asked which columns are features vs. labels. Example: if you are predicting “house price,” the price is the label; square footage, location, bedrooms are features.
Exam Tip: When you see “historical data with known outcomes,” that implies labeled data and supervised learning (training). When you see “predict for a new customer,” that implies inference. Don’t confuse “training dataset” with “test set”—the test set is held out for evaluation, not learning.
Common trap: assuming a model is the same as an algorithm. In exam wording, “choose an algorithm” is about the technique; “deploy a model” is about the trained result. Another trap: treating “data preprocessing” as optional—Azure ML workflows typically include data preparation steps, and questions may hint that poor data quality causes poor outcomes (a real-world and exam-relevant concept).
Supervised learning means you train using labeled data—examples where the correct answer is known. AI-900 most commonly tests two supervised tasks: classification and regression. Classification predicts a category (discrete label). Regression predicts a numeric value (continuous).
Classification patterns: “approve/deny a loan,” “spam/not spam,” “which product category,” “will the customer churn,” “detect fraud.” The output is a class label, sometimes with probabilities. Multi-class classification is still classification (more than two categories). A classic exam trap is “severity level 1–5”: although represented as numbers, it’s still classification if the numbers represent categories, not quantities.
Regression patterns: “predict sales,” “estimate delivery time,” “forecast energy consumption,” “predict temperature.” If the question emphasizes “amount,” “cost,” “time,” or “quantity,” regression is usually correct.
Exam Tip: Ask: “Would averaging two outputs make sense?” If yes, it leans regression (e.g., average predicted price). If no, it leans classification (e.g., average of ‘fraud’ and ‘not fraud’ is meaningless).
Supervised learning also introduces the idea of generalization: the model must perform well on unseen data. Overfitting (memorizing training patterns) and underfitting (too simple to learn signal) are frequently referenced indirectly, such as “performs well on training data but poorly in production.” Your best response is usually “overfitting” and/or “insufficient generalization,” not “bug in deployment.”
Common trap: confusing binary classification with anomaly detection. Fraud can be framed as either; if you have labeled fraud examples, classification fits. If you mostly have “normal” behavior and want to flag unusual cases with few labels, anomaly detection fits better (covered in Section 3.3).
Unsupervised learning uses unlabeled data. The exam’s primary unsupervised concept is clustering: grouping data points by similarity without pre-defined labels. If the scenario says “segment customers into groups,” “find similar documents,” or “group devices with similar telemetry,” clustering is the immediate association.
Clustering is often used as an exploration tool: you might cluster first, then assign business meanings to clusters (e.g., “budget shoppers,” “premium buyers”). The exam may test that clustering does not require a labeled outcome column—so if the prompt explicitly states “there is no historical outcome,” that’s a strong clue.
Anomaly detection is closely related but distinct: instead of forming groups, you identify rare or unusual observations. The exam expects you to recognize anomaly use cases like “detect credit card outliers,” “spot unusual network traffic,” “find manufacturing defects,” or “alert on sensor spikes.” Many anomaly solutions learn what “normal” looks like and flag deviations—useful when anomalies are rare and labeled examples are limited.
Exam Tip: If the goal is “create segments,” choose clustering. If the goal is “flag unusual/rare,” choose anomaly detection. Watch for wording like “outliers,” “rare events,” “deviates from baseline,” “unexpected pattern.”
Common trap: treating anomaly detection as “binary classification.” They can look similar, but classification implies you have labels for both classes (normal and anomalous). Anomaly detection is often used when you have mostly normal data and want to detect unknown or evolving anomalies.
Another trap: assuming unsupervised learning means “no evaluation.” You still evaluate clustering/anomaly models, but the metrics differ. For AI-900, focus on identifying the right approach rather than advanced cluster validity indices.
Evaluation questions on AI-900 are about choosing the right metric for the job and interpreting basic outcomes. For classification, the foundation is the confusion matrix, which counts true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN). From that, you get accuracy, precision, and recall.
Accuracy is the fraction of correct predictions overall. It is tempting—and often wrong—when classes are imbalanced (e.g., fraud is rare). A model can be 99% accurate by always predicting “not fraud,” but it is useless.
Precision answers: “When the model predicts positive, how often is it correct?” This matters when false positives are costly (e.g., blocking legitimate payments, sending unnecessary escalations). Recall answers: “Of all actual positives, how many did we catch?” This matters when missing a positive is costly (e.g., failing to detect fraud, missing a disease case).
Exam Tip: If the scenario emphasizes minimizing false positives, pick precision. If it emphasizes minimizing false negatives, pick recall. If it emphasizes overall correctness with balanced classes, accuracy can be acceptable.
For regression, a common metric is RMSE (root mean squared error), which measures the typical magnitude of prediction error in the same units as the target. Lower RMSE is better. If the question asks for “how close predictions are to actual numeric values,” RMSE (or related error measures) is the correct family.
Overfitting appears here too: if training performance is strong but test performance is weak, that indicates overfitting. A common exam trap is to suggest “add more features” as a remedy; often, the better conceptual answer is “use more training data,” “simplify the model,” or “regularize,” but at AI-900 level, simply recognizing overfitting/poor generalization is usually what’s tested.
Azure Machine Learning (Azure ML) is the Azure service for building, training, deploying, and managing ML models. AI-900 tests recognition of core components and when to choose Azure ML versus “prebuilt AI services.” If you need custom model training on your own data (especially for classic ML tasks), Azure ML is the go-to. If you need an out-of-the-box vision or language capability, Azure AI services may be simpler—Domain 2 focuses on Azure ML fundamentals, not the specialized APIs.
A workspace is the central container for Azure ML assets: datasets, models, experiments, endpoints, and related resources. If a question asks where you manage and organize ML resources, “workspace” is usually the answer.
Compute refers to the compute targets used for training or inference: compute instances (often for development), compute clusters (scalable training), and inference targets (endpoints). The exam often frames this as “scale training,” “run jobs on demand,” or “host a model for predictions.”
Pipelines orchestrate ML workflows as repeatable steps (data prep → train → evaluate → register → deploy). If the question emphasizes repeatability, automation, or CI/CD-like ML processes, pipelines are the concept being tested.
AutoML (automated machine learning) helps select algorithms and hyperparameters automatically. If a scenario says “we have limited ML expertise,” “quickly find the best model,” or “try multiple models automatically,” AutoML is the best fit.
Exam Tip: Watch for “custom model” versus “prebuilt.” If you must train with your own labels and choose algorithms/metrics, Azure ML is likely. If you just need OCR or sentiment analysis, the exam usually expects Azure AI services instead (covered in other domains).
Common trap: assuming Azure ML is required for all ML on Azure. The test sometimes presents Azure AI services as the simpler option. In Domain 2, however, when the story highlights model training, evaluation, deployment, and lifecycle management, Azure ML is the correct mental bucket.
This course uses timed simulations to build speed and accuracy. In Domain 2, your scoring gains come from fast scenario translation: identify (1) the ML type, (2) the phase (training vs. inference), (3) the metric, and (4) the Azure ML capability. Your timed caselets will often contain distractors that sound “more advanced” but are not required.
Rationale pattern 1—output type drives ML type: if the target is a category, it’s classification; if it’s a number, it’s regression; if there is no target and you are forming segments, it’s clustering; if the goal is “flag unusual,” it’s anomaly detection. Under time pressure, do not overthink the algorithm name; AI-900 prioritizes matching the task type.
Rationale pattern 2—metric aligns to business risk: accuracy is not the default when classes are imbalanced. Precision protects you from false positives; recall protects you from false negatives. RMSE belongs to regression. When the prompt mentions “rare events,” immediately question accuracy and look for precision/recall language cues.
Rationale pattern 3—lifecycle hints point to Azure ML features: “repeatable workflow” suggests pipelines; “limited expertise and want best model automatically” suggests AutoML; “scale training” suggests compute clusters; “central place to manage experiments and models” suggests workspace. If the scenario focuses only on consuming predictions, it’s inference and may involve a deployed endpoint.
Exam Tip: In timed sims, spend your first pass underlining (mentally) the noun phrases: “predict,” “estimate,” “group,” “outlier,” “evaluate,” “deploy,” “automate.” Those verbs usually map one-to-one to the correct option category.
Common traps in Domain 2 practice: (a) treating severity numbers as regression when they are categories; (b) picking accuracy for rare-event detection; (c) mixing up training and inference; (d) choosing Azure ML when the question is actually about a prebuilt AI service elsewhere. Your weak-spot analytics should track which trap you hit—then you drill that pattern until it becomes automatic.
1. A retail company wants to predict the number of units of a product it will sell next week based on price, promotions, and seasonality. Which machine learning type should you use?
2. You are building an Azure Machine Learning model to classify customer support tickets into categories (Billing, Technical, Account). Which items are labels in this scenario?
3. A banking app uses an existing trained model to score new loan applications submitted by customers in real time. Which stage of the ML lifecycle is occurring when the app generates the score for each new application?
4. A team trained a classification model and gets 99% accuracy on the training dataset but only 70% accuracy on a held-out test dataset. What is the most likely issue?
5. A company has historical data that includes customer demographics and whether each customer churned (Yes/No). They want the simplest Azure ML approach to automatically try algorithms and choose the best model without writing code. What should they use?
Domain 3 of AI-900 tests whether you can recognize common computer vision scenarios and map them to the right Azure service capability—quickly and confidently under time pressure. The exam is not looking for deep model-building knowledge; it is looking for correct service selection, correct feature identification (for example, “OCR” vs “image analysis”), and responsible-use awareness (especially around people and faces). This chapter follows a practical exam-coach flow: identify the vision task, map to Azure AI Vision capabilities, consider performance and responsible boundaries, then rehearse timed decision-making patterns.
In the timed sims portion of this course, vision-heavy items often hide the real requirement behind business wording like “extract text from photos of receipts” or “label the contents of product images.” Your job is to translate that wording into a workload type and output. When you do that translation first, the correct answer choices become much easier to eliminate.
Practice note for Identify vision tasks: image analysis, OCR, and detection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map tasks to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand performance considerations and responsible use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed simulation: vision-heavy items and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify vision tasks: image analysis, OCR, and detection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map tasks to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand performance considerations and responsible use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed simulation: vision-heavy items and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify vision tasks: image analysis, OCR, and detection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map tasks to Azure AI Vision capabilities: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand performance considerations and responsible use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Most AI-900 vision items reduce to three task families: (1) image analysis (describe/label what’s in an image), (2) optical character recognition (extract text), and (3) detection/recognition patterns (find things or people, sometimes with boundaries/locations). The exam commonly provides inputs such as a single image file, a set of images, a video frame, or a scanned document. The outputs, however, are what you should focus on: tags/captions, bounding boxes for objects, extracted text lines/words with coordinates, or structured fields (like invoice number) when document extraction is implied.
Start every question by asking: “Is the desired output words, labels, or locations?” Words usually means OCR. Labels/captions usually means image analysis. Locations (bounding boxes) could be object detection, people detection, or text coordinates from OCR. The exam likes to mix these signals: for example, “detect products and read the label text.” That is two tasks—image analysis plus OCR—often solvable with Azure AI Vision capabilities, but sometimes the presence of forms/invoices pushes you toward document-oriented extraction.
Exam Tip: Identify the “unit of work” and the “shape of output.” If the output is a JSON-like set of key-value fields (VendorName, Total, DueDate), that’s a document extraction pattern, not basic OCR alone.
Performance considerations that show up on AI-900 are typically high-level: latency (real-time vs batch), image quality, and the idea that poor lighting/angles reduce accuracy. If a scenario mentions mobile photos, glare, or skewed documents, expect OCR accuracy concerns and a recommendation to pre-process images or collect better inputs rather than assuming “a different SKU” fixes it.
Image analysis on Azure commonly means using Azure AI Vision to return descriptive information about an image. AI-900 expects you to recognize the difference between tags (keywords like “outdoor,” “car,” “tree”), captions (a sentence summarizing the scene), and objects (identified items, often with bounding boxes). If the scenario says “generate a description for accessibility,” that is a captioning use case. If it says “organize a photo library by categories,” tags are often the best match.
“People” and “background” cues can appear in subtle ways. “Count how many people are in the image” is still image analysis (detecting people as objects) unless the question explicitly asks to confirm identity. “Remove the background” or “identify the foreground object” is more about segmentation/background removal capabilities; the exam may phrase it as “separate a product from its background for a catalog.” You’re not expected to name model architectures—just match the scenario to a vision capability that can detect objects and/or separate background.
Common trap: confusing object detection with classification. Classification answers “what is in this image overall?” Detection answers “where are the instances of this object?” If a question includes phrases like “locate,” “bounding box,” “draw a rectangle,” or “highlight,” that is detection, not just tagging.
Exam Tip: When multiple answer choices include “Vision,” look for the verb: “analyze” (tags/captions), “detect” (objects/people and positions), “recognize text” (OCR). Verbs are often the fastest route to the correct option under timed conditions.
Responsible use also appears here: if the scenario involves people, ensure the requirement is content understanding (people present) rather than identity. AI-900 is sensitive to privacy and user consent topics when images contain people, and you should treat those as separate from generic image tagging.
OCR is the “extract text from images” workload, and AI-900 will test whether you can recognize it even when the prompt avoids the word “OCR.” Phrases like “read,” “extract,” “digitize,” “scan,” “convert to text,” and “searchable PDF” are OCR signals. Azure AI Vision includes OCR capabilities that return detected text, often with position data. The exam also distinguishes between printed versus handwritten text and between documents versus natural scenes.
Printed text (typed invoices, signs with block letters) is generally easier than handwriting. If the scenario emphasizes cursive notes, clinician handwriting, or handwritten forms, expect a mention of handwriting support and lower accuracy risks. “Documents vs scenes” is another frequent hinge: a photographed street sign is a scene; a scanned contract is a document. Both can require OCR, but documents often lead to structured extraction needs (for example, key-value pairs), while scenes focus on reading visible text reliably under varied lighting and angles.
Common trap: selecting “translation” or “language understanding” services when the prompt is purely about extracting text. Translation happens after OCR. The first step is always to get the text out.
Exam Tip: If the question mentions coordinates (line bounding boxes, word polygons) or “overlay the text position,” that strongly implies OCR output, not just “document processing.” Coordinate data helps highlight text back on the image.
Performance and quality considerations matter for OCR: skew, blur, low contrast, and compression artifacts reduce extraction quality. In practice, improving capture (lighting, focus, resolution) can outperform switching tools. On the exam, when you see constraints like “mobile camera images” and “near real-time,” pick the service that does OCR and then acknowledge that image quality and preprocessing are key performance factors.
AI-900 includes responsible AI fundamentals, and in vision this often shows up as questions involving faces, people, and identity. You need to distinguish between detecting a person/facial region (a vision task) and verifying or identifying a person (an identity-sensitive task). The exam may test whether you recognize that identity scenarios require stricter privacy, consent, and governance than generic image analysis.
Key principles you should apply: obtain informed consent, minimize data collection, protect biometric data, and provide transparency about how images are used. If the scenario is employee monitoring, customer recognition, or surveillance-like tracking, expect the “responsible use” angle to matter as much as the technical choice. You may need to recommend limiting scope (for example, count people rather than identify them) or ensuring human oversight.
Common trap: assuming “face detection” automatically implies “face recognition.” Detection is about finding a face region; recognition implies matching to a known identity, which is a different and more sensitive capability. If the requirement is “blur faces” or “detect faces to crop,” that is not an identity requirement.
Exam Tip: When an answer option mentions “identify a person” or “authenticate based on face,” pause and check whether the scenario explicitly states that identity confirmation is required and that consent and compliance requirements are addressed. If not, a safer non-identity approach (detect people, not who they are) is usually the better exam choice.
Safety boundaries also include preventing misuse. Vision solutions that can be used for profiling or discrimination should be governed carefully. On AI-900, demonstrating awareness—privacy by design, bias considerations, and restricted usage—is often enough to pick the best option among close distractors.
This is the highest-yield service-selection area in Domain 3: knowing when Azure AI Vision is sufficient and when the scenario is really “document extraction.” Azure AI Vision fits image analysis (tags, captions, object detection, people detection), and OCR for images. But if the prompt describes invoices, receipts, forms, contracts, IDs, or “extract fields into a database,” the workload is drifting toward document-oriented extraction, where the goal is not just text but structured information.
Document-oriented extraction concepts include key-value pair extraction, table extraction, and consistent field mapping (for example, InvoiceDate, TotalAmount). On AI-900, you may not need to name every product SKU, but you do need to recognize that “OCR alone” produces raw text, while document extraction produces structured outputs that align to business fields. If the question emphasizes downstream automation (populate ERP fields, validate totals, route approvals), it is likely testing this distinction.
Common trap: picking Azure AI Vision OCR for an invoice scenario that explicitly asks for “supplier name, invoice number, line items.” OCR can read the words, but it does not guarantee the semantic mapping to fields without additional processing. The exam expects you to spot the implied requirement for structure.
Exam Tip: Look for nouns like “fields,” “forms,” “tables,” “line items,” “key-value,” and verbs like “populate,” “extract structured data,” or “ingest into a system.” Those are your triggers to think beyond general OCR.
Performance considerations also differ. For generic image tagging, latency and API throughput may dominate. For documents, accuracy and consistency across templates matter, and you may need to consider that different layouts require different extraction strategies. Responsible use shows up here too: documents can contain PII. Minimization, encryption, access control, and retention policies are part of the “best answer” mindset even on a fundamentals exam.
In timed simulations, Domain 3 questions are often “two-step”: first identify the workload (image analysis vs OCR vs detection), then identify the capability/service that matches the output. To raise your score, practice eliminating distractors by naming the output you expect. If you can say, “This must return a caption,” you can immediately discard options that focus on text extraction or language modeling.
Rationale patterns you should internalize for exam readiness: (1) If the scenario says “describe” or “categorize images,” choose image analysis. (2) If it says “read text from images,” choose OCR. (3) If it says “locate objects/people,” choose detection with bounding boxes. (4) If it says “extract fields from invoices/forms,” think document-oriented extraction rather than raw OCR.
Common trap: overfitting to keywords in answer choices instead of scenario requirements. For example, seeing “Computer Vision” and selecting it automatically—without checking whether the scenario needs structured field extraction. Always translate scenario → output → capability, then match the answer.
Exam Tip: Manage time by using a 20-second rule on first pass: decide the workload type fast, pick the best match, and flag only if two answers both fit the same output. Most wrong answers fit a different output type entirely and can be eliminated quickly.
Weak-spot repair strategy: track which confusion happens to you most—(a) tags vs objects vs captions, (b) OCR vs document extraction, or (c) people detection vs identity. Then drill those contrasts with your own one-sentence “tell.” Example tells: “Captions are sentences; tags are keywords.” “OCR gives text; documents give fields.” “Detection finds; recognition identifies.” These small heuristics are exactly what you need to stay accurate under timed conditions in vision-heavy sets.
1. A retail company wants to extract the total amount and merchant name from photos of printed receipts taken on mobile devices. Which Azure capability should you use?
2. You have a folder of product images and need to automatically generate tags such as "shoe," "backpack," and "watch" to improve search. You do not need to train a custom model. Which Azure AI Vision feature best fits?
3. A manufacturing team wants to locate and draw bounding boxes around defects on parts (for example, scratches) in images. They have labeled images of defects and will iterate as new defect types appear. Which service is the best fit?
4. A company is building a kiosk that must verify whether a user is the same person as the ID photo they present. Which approach aligns best with AI-900 responsible AI expectations for vision scenarios?
5. You need to process thousands of images per minute to extract text. Latency and throughput are critical, and images may be slightly skewed or blurry. Which guidance best matches performance considerations for Azure AI Vision OCR?
This chapter targets AI-900 Domains 4: natural language processing (NLP) workloads and generative AI workloads on Azure. The exam expects you to recognize common language scenarios (classification, extraction, Q&A, conversation), map them to the right Azure services, and articulate the fundamentals of generative AI (LLMs, tokens, embeddings) with responsible use patterns. You are not being tested on building production architectures; you are being tested on selecting the best-fit capability and explaining why it fits the problem statement.
As you read, keep a key exam mindset: every question is usually a matching exercise between (1) the task type (extract vs classify vs generate vs retrieve), (2) the data type (free text, chat transcripts, documents), and (3) the Azure offering (Azure AI Language, bot concepts, Azure OpenAI). Tricky wording often appears as "needs deterministic answers" (Q&A) vs "needs creative or fluent generation" (generative AI), or "must return exact entities" (extraction) vs "must categorize documents" (classification).
We will build toward the chapter lessons: selecting NLP solutions for classification, extraction, Q&A, and conversation; understanding conversational building blocks; explaining prompting and grounding; and finishing with a timed-simulation mindset for mixed NLP + generative AI objectives.
Practice note for Select NLP solutions for classification, extraction, Q&A, and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand conversational AI building blocks and orchestration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts, prompting, and grounding on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed simulation: mixed NLP + generative AI objective questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Select NLP solutions for classification, extraction, Q&A, and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand conversational AI building blocks and orchestration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts, prompting, and grounding on Azure: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Timed simulation: mixed NLP + generative AI objective questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Select NLP solutions for classification, extraction, Q&A, and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand conversational AI building blocks and orchestration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 frequently frames NLP as a set of repeatable workload patterns. Your first job is to identify the pattern from the scenario language. The most common patterns are: sentiment analysis (opinion/attitude), key phrase extraction (topic clues), named entity recognition (people/places/organizations/dates), and summarization (condense text while preserving meaning). These are typically solved with prebuilt NLP capabilities rather than custom model training at the fundamentals level.
Sentiment analysis appears when the prompt mentions customer reviews, social posts, survey comments, "positive/negative/neutral," or prioritizing angry customers. Key phrase extraction shows up when the scenario says "identify main topics" or "generate tags" from text. Entity extraction (NER/PII detection) is indicated by mentions of extracting names, addresses, invoice numbers, dates, products, or redacting personal information. Summarization appears when users want a short synopsis of a long document, call transcript, or ticket history.
Exam Tip: Watch for the verb. If the task says "extract," it is usually entity/phrase extraction. If it says "classify," it is labeling into categories. If it says "generate" or "rewrite," that is generative AI (often Azure OpenAI), not classic NLP analytics.
A classic trap is to treat summarization as the same as key phrase extraction. Key phrases are sparse keywords; summarization is coherent text. Another trap: sentiment analysis is not the same as intent detection. Intent is a conversational concept (what the user wants), typically part of a bot/NLU flow, while sentiment is emotion/opinion. On the exam, you can often eliminate wrong answers by checking whether the output is structured (entities/phrases/sentiment scores) or free-form text (summaries, rewritten content).
When you see multiple requirements, prioritize the primary output. For example, "summarize and list key topics" could map to both summarization and key phrase extraction. In such cases, the exam typically expects the single best match to the main goal, unless explicitly asking for multiple features.
Azure AI Language is the umbrella service commonly associated with NLP tasks in AI-900. In exam terms, think of it as the place for text analytics features (sentiment, key phrases, entities, summarization) and for moving beyond prebuilt analysis into custom classification and custom extraction when the categories or entity types are specific to your business.
Text analytics capabilities are ideal when you can use general-purpose language understanding without training on your own labels. Scenarios: analyzing product reviews, extracting locations from news, finding key phrases in support tickets, detecting PII, or summarizing a meeting transcript. When the scenario mentions your own domain labelslike categorizing insurance claims into internal types, or extracting custom fields like "PolicyId" or "PartNumber" that arent standardyou should consider the custom options.
Custom text classification is used when the desired output is a label (or multiple labels) for a document or sentence, based on your examples. The exam may hint at this with phrases like "route tickets to the correct department," "assign a category," or "tag emails." Custom named entity recognition / extraction is used when you must pull out domain-specific fields from text, indicated by requirements such as "extract invoice IDs," "capture symptoms and medications," or "identify contract clause numbers."
Exam Tip: If the categories are stable and you can define them up front, classification fits. If the goal is to capture variable values embedded in text (IDs, names, amounts), extraction fits. A frequent trap is choosing classification when the scenario is clearly about extracting values.
At fundamentals level, you are not expected to know training pipelines, but you should recognize that custom approaches require labeled examples and iteration. Another common trap is mixing OCR with NLP: if the input is a scanned document image, you would first need OCR (computer vision) to convert it to text, and then use language features. AI-900 questions sometimes test whether you notice the input format constraint.
In weak-spot repair sessions, practice translating business phrasing into these building blocks: analyze (prebuilt), categorize (classification), extract fields (entity extraction), condense (summarization). This translation step alone improves accuracy under timed conditions.
Conversational AI questions in AI-900 usually test vocabulary and basic orchestration: bots, intents, utterances, entities (as conversational slots), and dialog flow. A bot is an application that users converse with via channels (web chat, Teams, etc.). An utterance is what the user says (a single message). An intent is the goal behind the utterance (e.g., "reset password," "track order"). A dialog or dialog flow is how the bot gathers information, confirms choices, and completes tasks across multiple turns.
Q&A workloads are different from open-ended generation. When the scenario emphasizes "answer from existing knowledge," "FAQ," "policy documents," or "provide consistent responses," you are in Q&A territory. The exam expects you to recognize that Q&A systems retrieve a relevant answer from a knowledge base rather than inventing one. This is why they are often preferred for compliance-sensitive or policy-driven responses.
Exam Tip: The fastest elimination rule: if the requirement says "must not hallucinate" or "answers must come from approved content," favor Q&A/retrieval patterns over generative-only chat.
Another tested concept is that conversational systems often require orchestration: detect intent, then route to either a Q&A response, a backend action (like creating a ticket), or a human handoff. Common traps include confusing sentiment with intent (emotion vs goal), or assuming that a bot always needs LLMs. Many bots are built with deterministic dialog flows and Q&A for reliability.
In timed questions, identify whether the conversation is task-oriented (collect info, execute action) or information-oriented (answer questions). Task-oriented flows emphasize dialog management and slot-filling; information-oriented flows emphasize Q&A retrieval. If both are present, the correct option is often the one that explicitly supports routing or integrates both patterns.
Generative AI on AI-900 is primarily conceptual: what large language models (LLMs) do, what tokens are, what embeddings represent, and when generation is appropriate. An LLM is trained to predict the next token and can produce fluent text, summarize, translate, classify, and more. Tokens are the chunks of text (not always words) that the model reads and generates; token limits matter for prompt size, conversation history, and cost. Embeddings are numeric vector representations of text (or other data) that capture semantic similarity, enabling search, clustering, and retrieval of relevant content.
Use generative AI when the output needs to be natural language that is synthesizeddrafting an email, rewriting content, summarizing in a particular style, brainstorming, or explaining code. Use embeddings-driven retrieval when the goal is to find relevant passages (semantic search) and then optionally use an LLM to compose an answer grounded in those passages.
Exam Tip: If the scenario says "find similar documents" or "semantic search" or "recommend related articles," think embeddings. If it says "write" or "compose" or "rephrase," think generation. If it says "answer using company documents," think retrieval + generation (grounding).
Common exam traps: (1) assuming LLMs are always the best choice for classification/extraction. Often, classic NLP analytics is cheaper, more predictable, and easier to validate. (2) ignoring the difference between retrieval and generation: embeddings do not generate text; they enable similarity matching. (3) overlooking token constraints: long documents may need chunking and retrieval rather than stuffing everything into the prompt.
Under time pressure, anchor on the required behavior: deterministic structured output (analytics/custom extraction) vs fluent synthesis (LLM). Then check constraints: must cite sources, must use internal data, must avoid fabrication. Those constraints push you toward grounding patterns rather than raw prompting.
Azure OpenAI Service provides access to OpenAI models hosted on Azure with enterprise controls. For AI-900, focus on what you do with it: chat/completions for text generation, embeddings for semantic similarity, and responsible deployment practices. When an exam question asks for "generate a draft," "summarize in a specific tone," or "create a conversational assistant," Azure OpenAI is a likely matchespecially when the desired output is open-ended natural language.
Prompting is the skill of instructing the model with clear context, constraints, and desired format. Good prompts specify role, task, inputs, output format, and guardrails (e.g., "If unknown, say you dont know"). Grounding (often via retrieval-augmented generation) reduces hallucinations by providing relevant source text for the model to use. The exam often signals grounding needs with phrases like "use company policies," "answer only from approved documents," or "include citations."
Exam Tip: When you see requirements about safety, accuracy, or approved sources, look for options mentioning grounding, retrieval, or using a data sourcenot just "use a larger model" or "increase temperature." Bigger models do not guarantee correctness.
Responsible generative AI is a high-frequency theme. You should be able to describe core concerns: harmful content, bias, privacy, security, and transparency. In scenario questions, note requests involving personal data, regulated industries, or end-user decision impacts. Your safest selection is typically the option that adds content filters, human review for high-stakes decisions, data minimization, and clear user disclosures.
Common traps include: (1) using generative AI as the sole authority for factual answers without verification, (2) forgetting that prompts and outputs can contain sensitive data, and (3) assuming that "temperature" is a safety control (it is primarily a creativity/variability control). For exam readiness, practice identifying the minimum responsible steps: restrict data, ground answers, monitor outputs, and provide a fallback path (e.g., escalate to a human) for uncertain cases.
This course is a mock-exam marathon, so your goal is not just knowledgeits speed and accuracy. For Domains 4, most missed questions come from misidentifying the workload type (analytics vs conversation vs generative) or overlooking a constraint (input is scanned, answers must be grounded, output must be structured). Use a repeatable decision checklist during timed sims.
Exam Tip: Watch for distractors that are "technically possible" but not the best fit. For example, an LLM can classify text, but if the question frames it as sentiment/key phrases/entities, Azure AI Language analytics is the intended answer.
Rationale patterns you should internalize: Sentiment/key phrases/entities/summarization map to Azure AI Language text analytics features. Domain-specific labels map to custom text classification. Domain-specific fields map to custom extraction/NER. FAQ-style consistent answers map to Q&A retrieval patterns. Multi-turn task completion maps to bot dialog concepts (intent, utterances, slot filling). Generative drafting and rewriting map to Azure OpenAI chat/completions. Semantic similarity and retrieval map to embeddings; grounded answers map to retrieval + generation rather than pure prompting.
For weak-spot analytics, tag every missed item with one of four causes: (1) misread verb (extract vs classify vs generate), (2) ignored constraint (grounding/safety), (3) confused services (Language vs OpenAI vs bot/Q&A), (4) rushed elimination. Then re-run a short set focusing only on that cause. Your score improves fastest when you reduce the same mistake class repeatedly, rather than re-reading everything.
1. A company wants to automatically route incoming support emails into categories such as Billing, Technical Issue, and Account Access. Which Azure capability is the best fit for this task?
2. You have unstructured customer messages and need to extract order numbers, delivery dates, and product names into structured fields for downstream processing. What should you use?
3. A team wants a chatbot that answers employee questions using content from internal policy documents. Answers must be based on the documents, and the bot should cite relevant passages. Which approach best matches this requirement?
4. You are designing a conversational solution on Azure. Which component is primarily responsible for interpreting a user's intent from natural language input (for example, determining whether the user wants to reset a password or check an order status)?
5. An application uses an LLM to generate responses. You want outputs to be more deterministic and reduce variation across runs for the same prompt. Which change is most appropriate?
This chapter is your capstone: two full mock runs (Part 1 and Part 2), a structured weak-spot analysis, and an exam-day checklist you can execute under pressure. AI-900 rewards candidates who can quickly recognize the workload type (vision vs. language vs. ML vs. generative AI), select the most appropriate Azure service or concept, and avoid “near-miss” options that sound plausible but don’t match the scenario. Your job in this chapter is not to “learn more content,” but to practice exam behavior: pacing, triage, and disciplined review.
Across both mock sets, you’ll deliberately rehearse how the exam phrases requirements (for example, “extract text from images” vs. “understand sentiment” vs. “build/train/deploy a model”), and how Responsible AI appears as constraints (fairness, transparency, privacy, reliability). You will also use weak-spot analytics to convert misses into a repeatable improvement plan. Keep your notes objective-tagged (Workloads, ML fundamentals, Vision, NLP, GenAI, Responsible AI, and exam readiness), because AI-900 grading is ultimately objective-based.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Your goal in the full mock is to mirror the real test environment: timed, uninterrupted, and decision-focused. Set a single block of time, remove reference materials, and commit to finishing even if you feel uncertain. The AI-900 exam tends to mix short conceptual items with scenario-based items; the trap is spending too long on a single scenario when a later question is easier and worth the same credit.
Use a three-pass pacing plan. Pass 1: answer everything you can immediately (your “fast wins”). Pass 2: return to flagged items and read more carefully for constraints like “must minimize labeling,” “real-time,” “no-code,” or “extract printed and handwritten text.” Pass 3: make final calls and eliminate remaining options decisively.
Exam Tip: Treat every question as an “identify the workload” exercise first. Ask: Is this prediction/classification/regression (ML), image analysis/OCR (Vision), text understanding/sentiment/key phrases (Language), conversation/QnA (Bot), or content generation (Azure OpenAI)? If you can label the workload in 5–10 seconds, you cut your time per question dramatically.
Common pacing trap: re-reading the scenario repeatedly without extracting the requirement. Instead, underline (mentally) the deliverable (what output is needed) and the constraint (latency, data type, responsible AI requirement, skill level). That is the minimum information needed to select the correct answer.
Mock Exam Part 1 (Set A) is domain-balanced: you should see a representative spread across AI workloads, ML fundamentals, Vision, NLP, and Generative AI on Azure, plus Responsible AI. While taking Set A, annotate each item with an objective tag as you go (or immediately after). Your objective is to verify breadth: you are checking whether you can correctly classify scenarios into the right domain quickly.
Focus areas that commonly separate passing from failing in a balanced set are service selection and ML vocabulary. For example, the exam expects you to distinguish “training a custom model” (Azure Machine Learning) from “using prebuilt AI capabilities via API” (Azure AI services such as Vision or Language). Another trap is confusing Computer Vision “image analysis” (tags, captions, objects) with OCR “read text,” or mixing Language tasks (sentiment, entity recognition, key phrases) with generative tasks (drafting, summarizing in a conversational style).
Exam Tip: For each scenario, ask “Is the model prebuilt or custom?” If the scenario implies you need to build from labeled data, tune features, or manage experiments, that leans toward Azure Machine Learning. If it implies calling an endpoint for common tasks (OCR, sentiment, translation), that leans toward Azure AI services.
Set A is also where you validate your “first instinct accuracy.” If your first pass score is low, your issue is likely domain identification or service mapping; if your second pass improves significantly, your issue is reading precision and trap avoidance.
Mock Exam Part 2 (Set B) is intentionally phrased in “hard-mode” exam wording: vague stakeholders, extra details, and distractor options that are technically true but not the best fit. Your task is to identify what the exam is actually testing—usually a single concept like OCR vs. image classification, supervised vs. unsupervised, or Azure OpenAI patterns vs. traditional NLP APIs.
Expect traps like these: (1) “analyze documents” where the hidden requirement is text extraction (OCR) rather than sentiment; (2) “predict a value” where options try to pull you into classification when it’s regression; (3) “create a chatbot” where the exam wants the concept of intent recognition and orchestration (Bot + language understanding concepts) rather than pure generative responses; (4) “generate marketing copy” where a candidate incorrectly selects Language analytics instead of Azure OpenAI.
Exam Tip: In hard-mode wording, locate the output noun. If the output is “a label/category,” think classification; “a number,” think regression; “groups,” think clustering; “text extracted,” think OCR; “a response in natural language,” think generative AI or conversational patterns.
Set B should feel harder even if the underlying content is the same. If you miss questions here, it often indicates you are reacting to familiar keywords instead of reading for deliverables and constraints. Your review should focus on why the wrong option was tempting, and what wording would have disqualified it.
Your weak spot analysis should be mechanical, not emotional. Build an error log with three columns: (1) objective tag (e.g., Vision-OCR, ML-Types, GenAI-Responsible Use), (2) failure mode, and (3) corrected rule. Failure modes are typically: misread requirement, confused service, confused ML type, or “fell for distractor.” The corrected rule must be short enough to recall during the exam.
Exam Tip: Don’t just write “study OCR.” Write a decision rule: “If the output is text from an image/PDF scan, choose OCR/Read; image analysis is for objects/tags/captions.” Decision rules translate directly into points.
When you redo, focus on speed and certainty. If you need more than a short recall to justify the answer, your understanding is still fragile. The goal is automaticity: see the scenario, label the workload, pick the service/concept, move on.
Finally, quantify progress by objective: your target is not just a higher overall score, but fewer misses in your worst two domains. That’s the fastest way to raise your passing probability.
This recap is a conceptual cheat sheet to use during your final review. AI-900 rarely requires deep implementation detail; it expects correct matching between scenario and capability. Build a mental map from “what is needed” to “what Azure offers.”
Exam Tip: If two options both “can work,” choose the one that is most direct and managed for the task described. The exam favors the simplest service that satisfies the requirement without unnecessary custom building.
Common trap: interpreting “AI” as always meaning “train a model.” Many AI-900 scenarios are solved with prebuilt APIs; only select custom training when the scenario demands it (unique labels, specialized domain, or control over the model lifecycle).
On exam day, your performance is a function of pacing and composure as much as knowledge. Start with a quick scan mindset: you are not trying to prove you’re right with a dissertation—you’re matching requirements to the best-fit concept or service. Use the same three-pass strategy from your timed simulations.
Exam Tip: If you are stuck, force an elimination. Ask: “Which option is least aligned with the required output?” Removing one wrong option reduces cognitive load and improves guessing accuracy even when you remain uncertain.
Common exam-day trap: changing correct answers late due to anxiety. Only change an answer if you can state a clear rule that contradicts your original choice (for example, you realized it’s OCR because the output is extracted text, not image labels). If you can’t articulate the rule, keep your first selection.
Finish with a final checklist: confirm you answered every question, revisit only flagged items, and ensure you didn’t misread “best” vs. “possible.” This chapter’s full mock runs and weak-spot repair process are designed to make exam day feel familiar—and therefore manageable.
1. A retail company wants to build a solution that can read product labels from photos taken in a warehouse and return the extracted text to an inventory system. Which Azure AI service should you use?
2. You are reviewing a practice exam result and notice you frequently confuse workloads such as language understanding vs. document text extraction. Which action is the best weak-spot repair approach for AI-900 based on objective-based study?
3. A team is building a customer support chatbot. They want the bot to generate natural-language responses and also cite the source policy documents used to create the answer. Which approach best aligns with this requirement?
4. A bank is deploying an AI model that helps decide whether to approve loan applications. The compliance team requires the solution to provide explanations for decisions and ensure protected groups are not disproportionately denied. Which Responsible AI principles are most directly addressed?
5. During the exam you have 12 questions remaining and 10 minutes left. Several questions are long scenario items with similar-sounding service options. What is the most appropriate exam-day strategy to maximize your score?