AI Certification Exam Prep — Beginner
Master AI-900 concepts and practice what Microsoft tests—without coding.
This course is a complete, beginner-friendly prep path for the Microsoft AI-900: Azure AI Fundamentals certification exam. It’s designed for non-technical professionals who want to confidently describe AI concepts, recognize the right Azure AI services for common scenarios, and answer Microsoft-style questions under exam conditions—without requiring coding or data science experience.
The AI-900 exam evaluates your understanding across these official domains:
Throughout the course, you’ll practice selecting the right approach and service for realistic workplace scenarios—exactly the kind of decision-making the exam tests.
The course is structured like a focused exam-prep book with six chapters:
Many AI-900 learners know the buzzwords but struggle with what the exam actually asks: choosing between AI workload types, interpreting basic ML concepts, and matching Azure services to business needs. This course closes that gap by:
If you’re ready to start preparing, set up your learning access and schedule your study plan. You can Register free to begin, or browse all courses to compare additional certification paths.
By the end of this course, you’ll be able to describe AI workloads clearly, explain foundational ML principles on Azure, identify the right Azure services for computer vision and NLP scenarios, and articulate generative AI concepts and responsible usage—then prove it with a timed mock exam aligned to the official AI-900 domains.
Microsoft Certified Trainer (MCT) | Azure AI Fundamentals Specialist
Jordan McAllister is a Microsoft Certified Trainer who helps beginners pass Microsoft fundamentals exams through clear, business-friendly explanations. Jordan has supported learners preparing for Azure and AI certifications with focused practice aligned to official objectives.
AI-900 is designed for non-technical professionals who need to speak confidently about AI concepts and the Azure services that implement them. This chapter orients you to what the exam measures, how to register and sit the exam, how Microsoft writes questions, and how to build a realistic plan that tracks your readiness by domain. You will also set expectations: AI-900 rewards accurate concept recognition and good service selection more than hands-on coding. That said, a lightweight Azure learning environment can be a force-multiplier—especially for understanding the difference between “what AI can do” and “which Azure product fits the scenario.”
Throughout this course, you will map every topic back to the outcomes: describing AI workloads and solution scenarios (including responsible AI), explaining machine learning fundamentals and when to use Azure Machine Learning, identifying computer vision and NLP workloads and the right Azure AI services, and describing generative AI workloads and Azure OpenAI capabilities responsibly. Start building the habit now: for any prompt, ask (1) what workload is this, (2) what service family fits, and (3) what responsible AI consideration should be mentioned.
Exam Tip: AI-900 questions often look like “business language.” Your advantage is translating business goals into a workload type (vision, language, ML, generative AI) and then choosing the most appropriate Azure service category—without overengineering.
Practice note for Understand the AI-900 exam format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set up your test environment (online or test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a realistic study plan and track your readiness by domain: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Learn how to approach Microsoft-style questions and eliminate distractors: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Set up a lightweight Azure learning environment (optional) and resource checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand the AI-900 exam format, timing, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set up your test environment (online or test center): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a realistic study plan and track your readiness by domain: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Learn how to approach Microsoft-style questions and eliminate distractors: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 tests foundational understanding rather than implementation details. The exam objectives are published by Microsoft and periodically updated, so your first step is to download the current skills outline and treat it as your checklist. The typical domains include: describing AI workloads and considerations (including responsible AI), describing fundamental principles of machine learning on Azure, and describing features of computer vision, natural language processing, and generative AI workloads on Azure.
Your “weighting mindset” matters even if you don’t memorize percentages. Allocate more study time to topics that produce many scenario questions: service selection (which Azure AI service fits), basic ML terminology (training vs. inference, classification vs. regression), and responsible AI principles. Lighter areas (like some process definitions) still matter, but they are usually tested as supporting details within scenario prompts rather than as standalone trivia.
What the exam is really measuring is whether you can: (1) recognize the workload (vision, language, generative AI, predictive ML), (2) pick the correct Azure service family (Azure AI Vision, Azure AI Language, Azure Machine Learning, Azure OpenAI), and (3) articulate basic considerations (data, accuracy, bias, privacy). Common traps include choosing a “powerful-sounding” option that is not the most direct fit, or confusing platform services (Azure Machine Learning) with prebuilt AI services (Azure AI services). Another trap is mixing up “model training” tools with “consuming a pretrained model.”
Exam Tip: When two answers look plausible, look for the one that most directly satisfies the scenario with the least custom work. AI-900 generally favors prebuilt services (Vision/Language/OpenAI) when the prompt describes common tasks like OCR, sentiment analysis, or chat-based summarization.
Logistics are easy to underestimate, and candidates lose points from avoidable stress. Register through Microsoft Learn or the certification dashboard and schedule with Microsoft’s exam delivery partner (often Pearson VUE). You can typically choose an online proctored exam or a test center appointment. Pricing varies by region and discounts (student, employer, event vouchers) may apply, so verify current cost before you commit.
For registration, ensure your legal name matches your government-issued ID. ID mismatch is a classic last-minute failure point. If you need accommodations, request them early through Microsoft’s accommodations process; approvals can take time and may require documentation. Understand policies: rescheduling windows, retake rules, and what counts as a violation during online proctoring.
Online testing requires a compliant environment: stable internet, a quiet room, desk cleared of papers, and often a webcam/microphone check. The proctor may ask you to show your desk and room; having a second monitor, phone, or notes nearby can trigger warnings or termination. For test centers, arrive early and plan for check-in time, lockers, and security procedures.
Exam Tip: Do a “test-run” of your online setup 24–48 hours before exam day: system check tool, webcam angle, lighting, and network stability. Treat environment issues as a study topic—because they can stop you from taking the exam at all.
AI-900 uses scaled scoring; the passing score is commonly published as 700 on a 1000-point scale, but the number of questions and exact scoring behavior can vary. Do not assume that every question is weighted equally. Some items may be unscored (used to evaluate future questions). Your goal is consistency across domains rather than perfection in one area.
Expect Microsoft-style formats such as multiple-choice (single answer), multiple response (“choose all that apply”), drag-and-drop matching (for concepts, services, or workflow steps), and scenario-based questions. Case study questions may present a longer business story, requirements, and then multiple questions tied to the same scenario. Read carefully: later questions may introduce new constraints (budget, compliance, latency) that change the best service choice.
Common traps include missing key words like “prebuilt,” “custom,” “real-time,” “on-device,” or “explainability.” Another frequent distractor is a correct statement that does not answer the question asked—for example, selecting Azure Machine Learning when the prompt clearly wants a prebuilt OCR feature (Azure AI Vision) or selecting a database/analytics service when the question is about model inference.
Exam Tip: For multiple-response items, don’t “average” your way to safety. Verify each selected option independently against the scenario requirements. A single wrong selection can cost you the whole item depending on scoring rules.
Time management is usually manageable, but don’t rush the first third of the exam. Establish a pacing rule: read the question stem, identify the workload and constraint, eliminate two distractors, then choose. If you are unsure, mark and move on—momentum prevents costly panic.
If you are new to AI and Azure, your biggest risk is “vocabulary overload.” The cure is a domain-mapped study plan with spaced repetition. Start by creating five buckets aligned to the outcomes: AI workloads/responsible AI, ML fundamentals/Azure Machine Learning, computer vision/Azure AI Vision, NLP/Azure AI Language, and generative AI/Azure OpenAI. Every study session should touch at least two buckets, and every week should revisit all five.
Spaced repetition means you re-encounter the same concepts repeatedly over increasing intervals (for example: Day 1, Day 3, Day 7, Day 14). This is particularly effective for service selection (which service does OCR? which does entity recognition? which is used for training vs. consumption?). Do not rely on a single long weekend of cramming—AI-900 questions are designed to test recognition under light ambiguity, which requires memory consolidation.
Build a simple readiness tracker. For each domain, rate your confidence (e.g., 1–5) on: definitions, service identification, and scenario selection. After practice, update your rating based on mistakes. Your study plan should follow data: if you consistently confuse Azure Machine Learning with Azure AI services, add a focused session on “custom ML vs. prebuilt AI.”
Exam Tip: Learn the “decision shortcut”: prebuilt service first (Vision/Language/OpenAI) unless the scenario explicitly requires custom model training, specialized datasets, or MLOps—then Azure Machine Learning becomes the likely answer.
Finally, consider setting up a lightweight Azure learning environment (optional). Even without coding, exploring the Azure portal and reading service descriptions makes the exam’s product names feel real, reducing cognitive load on test day.
Responsible AI is not a separate “ethics-only” segment; it appears across scenarios in every domain. Microsoft commonly frames responsible AI around principles such as fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability. On AI-900, you will not be asked to write policies, but you will be expected to recognize risks and select the best mitigation or best practice.
In ML contexts, responsible AI shows up as questions about bias in training data, explainability (being able to describe why a model made a prediction), and monitoring drift over time. In computer vision, risks can include demographic bias in facial analysis scenarios or privacy concerns with surveillance-like use cases. In NLP, watch for data leakage (sensitive customer text), toxic content, or misinterpretation of sentiment across cultures or languages. In generative AI, expect questions about groundedness (avoiding hallucinations), content filtering, prompt injection risks, and human oversight.
A common exam trap is choosing an answer that improves accuracy but ignores governance. For example, “collect more personal data” might increase performance but conflicts with privacy principles unless justified and protected. Another trap is assuming responsible AI is only about “being nice” rather than being systematic: documentation, access controls, auditability, and clear communication of limitations are all responsible AI actions.
Exam Tip: When you see words like “sensitive,” “regulated,” “public-facing,” “high impact,” or “customers,” mentally append a responsible AI checklist: privacy, bias, transparency, and human review. One of the options will usually align directly to these controls.
Your practice workflow should be lightweight but disciplined. Start with a single page of “service-to-task” notes: map common tasks (OCR, image tagging, sentiment analysis, entity extraction, translation, chat completion, embeddings) to their Azure service families. Keep it simple and update it as you learn. The goal is fast recall under time pressure, not encyclopedic detail.
Use flashcards for high-yield distinctions: training vs. inference, supervised vs. unsupervised learning, classification vs. regression vs. clustering, and which scenarios call for Azure Machine Learning versus Azure AI services. Spaced repetition tools work well here, but paper cards are fine if you review them on a schedule.
Then apply a review loop: do a set of practice questions, tag every miss by domain and by mistake type (concept gap, misread stem, fell for distractor, or overthought). Your next study session should target the top two mistake types. This is how you “track readiness by domain” without guessing.
Exam Tip: When reviewing mistakes, rewrite the question in your own words as a workload statement: “This is NLP + sentiment analysis” or “This is vision + OCR.” If you can label the workload correctly, you can usually eliminate at least two options immediately.
Finally, if you set up an Azure learning environment, keep a resource checklist to avoid surprise costs: use free tiers where available, set budgets/alerts, delete test resources, and prefer guided sandboxes (like Microsoft Learn modules) when possible. The objective is familiarity, not building production solutions.
1. You are advising a marketing manager who is preparing for AI-900. The manager asks what the exam is primarily designed to validate. Which statement best describes the focus of AI-900?
2. A company wants employees to take AI-900 from home. They ask what is most important to confirm when setting up the test environment for an online proctored exam. What should you emphasize first?
3. You are creating a study plan for AI-900 and want to track readiness effectively. Which approach best aligns with AI-900 preparation strategy described in the chapter?
4. A question on the AI-900 exam describes this scenario: "A retailer wants to automatically categorize product photos and detect whether images contain a logo." What is the best first step to answer Microsoft-style questions like this?
5. A non-technical team wants a lightweight way to learn Azure AI concepts while preparing for AI-900. Which statement best describes the role of setting up a basic Azure learning environment for this exam?
This chapter targets the AI-900 objective area that asks you to describe AI workloads and common AI solution scenarios. On the exam, you are rarely asked to build anything; you are asked to recognize the workload type (vision, language, forecasting, knowledge mining, or generative AI), understand basic machine learning (ML) terms, and apply responsible AI principles to realistic workplace contexts. Your job is to read a scenario, identify what the system must do (classify, predict, extract, generate), then map that to the right AI approach and Azure capability family.
A common candidate mistake is to treat “AI” as one tool. AI-900 expects you to differentiate AI vs ML vs deep learning in plain language: AI is the broad umbrella for systems that appear to perform intelligent tasks; machine learning is a subset where models learn patterns from data; deep learning is a subset of ML that uses multi-layer neural networks, often strong for vision, speech, and language. In exam scenarios, deep learning is usually implied (not named) when you see images, speech, or large-scale language understanding, but you still answer at the workload level unless the question explicitly asks about model types.
Throughout this chapter, focus on two exam habits: (1) translate business language into ML vocabulary (features, labels, training, inference), and (2) spot keywords that signal workload type (e.g., “detect objects in photos” → vision; “summarize emails” → generative AI + language; “search PDFs for answers” → knowledge mining). Exam Tip: When two options seem plausible, pick the one that matches the output the scenario needs (prediction, extraction, generation), not the input format.
Practice note for Differentiate AI, machine learning, and deep learning in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match real business problems to AI workload types (vision, language, forecasting): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain key AI concepts: features, labels, training vs inference, and model drift: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles to common workplace scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Complete domain-focused exam-style drills for 'Describe AI workloads': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate AI, machine learning, and deep learning in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Match real business problems to AI workload types (vision, language, forecasting): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain key AI concepts: features, labels, training vs inference, and model drift: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 tests whether you can match a business problem to an AI workload category. Start by asking: “What kind of input do we have?” and “What kind of output do we need?” For vision workloads, the input is images or video and the output might be tags, detected objects, text extracted from images (OCR), or descriptions. Typical workplace examples include identifying damaged inventory from photos, reading license plates, or checking whether a factory worker is wearing safety gear.
Language workloads focus on understanding or producing meaning from text (and sometimes speech-to-text as a front door). Outputs include sentiment, key phrases, named entities (people, places, organizations), classification (routing support tickets), or translation. AI-900 often describes these as “natural language processing (NLP)” scenarios. Don’t confuse this with generative AI: traditional NLP typically extracts or classifies; generative AI produces new text.
Knowledge mining sits between content and search. The goal is to index and retrieve insights from large collections of documents—PDFs, scans, transcripts—so users can search, filter, and discover relationships. In practice, solutions use OCR plus language extraction and then store the enriched index for search. Exam Tip: If a scenario emphasizes “search across thousands of documents” or “build an index with extracted entities,” think knowledge mining rather than plain NLP.
Generative AI creates content: drafts, summaries, Q&A responses, or code-like text. It is best recognized by verbs like “generate,” “summarize,” “rewrite,” “compose,” or “chat.” On Azure, this points you toward Azure OpenAI capabilities. A classic exam trap is assuming generative AI is required whenever text is involved; if the task is to label, classify, or extract fields, traditional language services (or vision OCR) may be a better fit.
Business mapping skill: “forecasting sales” is also an AI workload (prediction/regression) even though it isn’t vision or language. On AI-900, this typically appears as ML for numeric prediction (time series–like), not as generative AI.
Core ML vocabulary appears constantly in AI-900 questions. A dataset is the collection of examples you use for learning or evaluation. A feature is an input attribute used to make a prediction (e.g., “account age,” “purchase history,” “image pixels,” “number of prior incidents”). A label is the correct answer for supervised learning (e.g., “fraud/not fraud,” “defective/not defective,” or a numeric value like “days to delivery”).
Training is when the algorithm learns patterns from a dataset. Inference is when the trained model is used to make predictions on new data. On the exam, training is associated with building or updating a model; inference is associated with deploying and using it in an app or workflow. Exam Tip: If the scenario says “use past data to build a model,” that’s training; if it says “predict for each new customer order,” that’s inference.
Understanding training vs inference helps you interpret Azure services questions. Azure Machine Learning is often referenced as the platform used to train, evaluate, and manage ML models. Many Azure AI services (Vision, Language) provide prebuilt models where you typically focus on inference, configuration, and responsible use rather than building the algorithm from scratch.
Model drift is another testable idea: performance can degrade over time because the real world changes (customer behavior shifts, new product categories appear, a camera’s lighting changes). Drift is not “the model forgot”; it is “the data distribution changed.” The practical response is monitoring, retraining, and validating with fresh data. A subtle trap: drift can happen even if your code is unchanged. If a scenario mentions “accuracy has dropped over the last quarter,” think drift and the need to review training data and model monitoring.
When you see “features” and “labels” in answer choices, you are being tested on your ability to map business columns to ML roles. A column can be a feature in one problem and a label in another depending on what you are trying to predict.
AI-900 expects you to speak in outcome terms: “How do we know the model is good enough?” The simplest metric is accuracy: the percentage of correct predictions. Accuracy is useful when classes are balanced (roughly equal numbers of each outcome). However, many business problems are imbalanced (fraud is rare, equipment failure is rare). In those cases, accuracy can be misleading: a model can be 99% accurate by always predicting “not fraud.”
That’s why you must recognize precision and recall. Precision answers: “When the model predicts positive, how often is it correct?” Recall answers: “Of all actual positives, how many did we catch?” These map directly to workplace risk: a medical screening system often prioritizes recall (don’t miss cases), while a high-cost investigation process may prioritize precision (don’t waste analyst time). Exam Tip: When a scenario emphasizes “minimize false alarms,” lean toward precision; when it emphasizes “don’t miss any real cases,” lean toward recall.
To reason about these metrics, you also need error types. A false positive is predicting something is true when it is not (flagging a legitimate transaction as fraud). A false negative is predicting something is false when it is true (missing a fraudulent transaction). The exam commonly frames this as “which error is more costly?” Your answer should align to business impact, not general ML theory.
For regression/forecasting (predicting a number), the exam may describe “difference between predicted and actual” as error. You don’t need advanced statistics, but you should recognize that lower error means better predictions, and that evaluation must happen on data not used for training (to avoid overestimating performance).
Common trap: thinking one metric is always best. AI-900 wants you to choose the metric that matches the scenario’s cost of mistakes.
Responsible AI shows up as both direct questions and as “best choice” decision-making in scenarios. Four principles you must be able to apply in plain workplace terms are fairness, reliability and safety, privacy and security, and transparency.
Fairness is about avoiding unequal harmful outcomes across groups. For example, a hiring screening model trained on historical decisions could learn biased patterns and disadvantage certain applicants. The exam often tests whether you can identify this risk even when it is not explicitly called “bias.” Exam Tip: If the scenario involves people (hiring, lending, healthcare), expect fairness to be relevant and choose actions like evaluating performance across demographics and using representative data.
Reliability and safety means the system behaves consistently and fails gracefully. In AI terms, this includes robust performance under changing conditions and appropriate handling of uncertain predictions. If a model is used in a safety-critical environment, you should expect monitoring, human review, and fallback procedures. This pairs naturally with model drift: monitoring is both a reliability practice and a maintenance necessity.
Privacy and security is about protecting sensitive data used in training and inference. Scenarios might mention customer PII, medical records, or confidential documents. You are being tested on recognizing that AI solutions should follow least privilege, data minimization, and secure storage/processing practices. In generative AI scenarios, privacy includes preventing accidental leakage of sensitive prompts or outputs.
Transparency means stakeholders understand that AI is being used and can interpret outcomes appropriately. It includes explainability (“why did the model reject this loan?”), documentation, and communicating limitations. A frequent trap is choosing “make the model more complex” as a fix; transparency often improves when you add explanations, clear user messaging, and governance—not just accuracy.
When in doubt, pick the answer that adds monitoring, evaluation across groups, and clear communication—these are “safe defaults” aligned with AI-900 expectations.
AI-900 frequently asks you to select the most appropriate approach rather than a specific algorithm. Think of three buckets: rules-based logic, machine learning, and generative AI. Rules-based solutions are best when the logic is stable, explicit, and easy to write (e.g., “if invoice amount > $10,000 then require approval”). ML is best when patterns are too complex for explicit rules but you have historical data to learn from (fraud detection, demand forecasting, defect classification). Generative AI is best when you need flexible language generation or transformation (drafting, summarizing, conversational Q&A).
A key exam skill is identifying when ML is unnecessary. If a scenario is deterministic and unlikely to change, rules may be cheaper, easier to audit, and more transparent. Conversely, if the scenario includes “too many combinations,” “difficult to define rules,” or “needs to improve with data,” that points to ML. Exam Tip: Phrases like “learn from historical outcomes” and “predict” are ML signals; phrases like “compose,” “summarize,” and “rewrite” are generative AI signals.
Also distinguish prebuilt AI services vs custom ML. If the task is common and standardized (OCR, entity extraction, image tagging), prebuilt Azure AI services are usually the best match. If the task is unique to your organization (predicting churn from internal usage patterns), Azure Machine Learning is a typical fit for training custom models. The exam likes pragmatic choices: use prebuilt when it meets requirements; build custom when domain specificity demands it.
Generative AI introduces new considerations: factuality (hallucinations), policy compliance, and data protection. Even if generative AI can do a task, it may not be the best choice for structured extraction where deterministic output format matters. In those cases, traditional language extraction or a constrained ML classifier may outperform in reliability and auditability.
On AI-900, “best approach” answers usually balance capability with risk and operational simplicity.
This section prepares you for the domain-focused “Describe AI workloads” items without turning into a quiz. The exam pattern is consistent: you receive a short business scenario, then you must (1) identify the workload type, (2) identify whether it is training or inference, and (3) consider a responsible AI concern. Build a quick mental checklist you run in under 15 seconds.
Step 1: Identify the artifact. Images/video → vision. Text documents/emails/chats → language. “Search across documents and extract insights” → knowledge mining. “Generate a draft/summary/answer” → generative AI. Numeric forecasting (“predict next month’s demand”) → ML regression/forecasting.
Step 2: Identify the action verb. “Classify,” “detect,” “extract,” “translate,” “route,” “predict” usually indicate non-generative workloads. “Compose,” “summarize,” “rewrite,” “chat,” “create” usually indicate generative AI. Exam Tip: If you see “extract key phrases/entities,” do not pick generative AI just because text is involved—this is classic NLP extraction.
Step 3: Map to ML vocabulary. If the scenario mentions a historical outcome column (approved/denied, defective/not, dollars spent), that’s a label and implies supervised learning. If it says “group similar customers,” that’s clustering (unsupervised). If it says “use the model in an app to score new requests,” that’s inference. If it says “retrain monthly,” that’s training plus drift management.
Step 4: Choose the metric and watch for traps. If positives are rare and missing them is costly, recall matters. If false alarms are costly, precision matters. If the classes are balanced and the cost of each error is similar, accuracy may be sufficient. Be ready to justify error types: false positives vs false negatives, and which one the business can tolerate.
Step 5: Apply responsible AI. For people-impacting decisions, fairness and transparency are almost always relevant. For sensitive data, privacy and security are relevant. For changing environments, reliability and monitoring (including drift) are relevant. Many exam options include “increase training data” as a generic fix; prefer answers that directly address the stated risk (e.g., evaluate across demographics for fairness; add monitoring for drift; restrict access and protect PII for privacy).
Use these steps to turn every scenario into a structured decision: workload category → approach (rules/ML/generative) → training vs inference → metric → responsible AI check. That is exactly what this AI-900 domain expects you to demonstrate.
1. A retail company wants to automatically detect whether product photos contain a visible barcode and, if present, draw a bounding box around it. Which AI workload type best fits this requirement?
2. A manager says, "We want AI to approve or deny loan applications." You are translating the requirement into ML terms for a classification model. Which pairing correctly identifies the labels and features?
3. A contact center wants an AI solution that can read incoming customer emails and generate a short summary plus a suggested reply draft for an agent to review. Which workload type is the best match?
4. A manufacturer trained a model last year to predict equipment failures from sensor readings. Recently, the model’s accuracy dropped because new machines were installed and the sensor patterns changed. What concept best describes this issue?
5. An HR team wants to use an AI model to screen resumes and rank candidates. Which action best supports the responsible AI principle of fairness?
This chapter maps directly to the AI-900 exam objective area that asks you to explain fundamental machine learning (ML) principles on Azure and when to use Azure Machine Learning. The exam is not looking for coding skill; it’s checking whether you can identify the right ML approach for a business problem, describe how models are trained and evaluated in plain language, and choose the right Azure Machine Learning workflow (Automated ML, designer, or code-first) at a conceptual level.
On the test, many questions are disguised as business scenarios: “predict next month’s demand,” “classify emails,” “group customers,” “detect unusual transactions.” Your job is to recognize the ML problem type and then connect it to how Azure Machine Learning helps you build, train, evaluate, and deploy a model.
As you read, keep two recurring exam patterns in mind: (1) the exam loves to test the difference between prediction (supervised learning like regression/classification) and pattern discovery (unsupervised learning like clustering), and (2) the exam frequently tests your ability to explain overfitting and why validation matters—without using heavy math.
Practice note for Identify regression, classification, clustering, and time-series use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain training/validation concepts and overfitting in non-technical terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose between Automated ML, designer, and code-first workflows conceptually: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Describe the Azure Machine Learning workspace and core components at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Complete exam-style questions for 'Fundamental principles of ML on Azure': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify regression, classification, clustering, and time-series use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain training/validation concepts and overfitting in non-technical terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose between Automated ML, designer, and code-first workflows conceptually: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Describe the Azure Machine Learning workspace and core components at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Complete exam-style questions for 'Fundamental principles of ML on Azure': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to identify common ML workloads and map them to the right problem type. The quickest way is to look at the output you want. If the output is a category, think classification. If the output is a number, think regression. If there are no labels and you want to discover natural groupings, think clustering. If the key requirement is predicting future values over time (often with seasonality), think forecasting (time-series).
Classification predicts which bucket an item belongs to: spam vs. not spam, loan approved vs. denied, “high/medium/low” risk. The exam often includes language like “assign,” “categorize,” “label,” or “detect whether.” Regression predicts a continuous value: house price, energy usage, delivery time in minutes. Look for “estimate,” “predict the amount,” “forecast a value” (but not necessarily time-series).
Clustering groups similar items without pre-defined labels: segment customers by purchasing behavior or group devices by telemetry patterns. You’ll see wording like “group,” “segment,” “discover patterns,” or “identify cohorts.” Forecasting is a specialized type of regression where time order matters: demand next week, call volume next hour, inventory next quarter. Exam scenarios may include seasonality, trends, or repeated time intervals.
When choosing between regression and forecasting on the exam, pay attention to time. Predicting “next month’s sales” implies time-series. Predicting “sales based on ad spend” could be regression if time isn’t central. The correct answer hinges on whether the ordering of observations (time) is essential to the model.
AI-900 frequently tests training and validation concepts using everyday language. The key idea: a model must perform well not only on the data it learned from, but also on new, unseen data. This is called generalization. To check generalization, we split data into at least a training set (what the model learns from) and a test set (what we hold back to evaluate final performance). Often, there is also a validation set used during model selection or tuning.
In non-technical terms, the training set is like practice questions, and the test set is like the final exam. If you only practice the exact same questions and memorize answers, you may score high in practice but fail on new questions. That’s overfitting: the model “memorizes” quirks and noise in training data instead of learning reliable patterns.
The exam also expects you to know that more complex models can overfit more easily, and that good evaluation needs data the model hasn’t seen. If a scenario says, “The model performs very well on training data but poorly in production,” that’s a classic overfitting clue.
Another term you may see is underfitting, where the model is too simple to capture patterns, leading to poor performance even on training data. For AI-900, you typically just need to recognize overfitting vs. underfitting at a conceptual level and understand why data splitting and validation protect against misleading results.
Even though AI-900 is non-technical, it still tests the principle that models are only as good as the data and features provided. A feature is an input signal the model uses to make predictions—like age, location, purchase history, device temperature, or time of day. Feature engineering means choosing, transforming, or creating features to better represent the real-world problem.
Data quality issues show up on the exam as missing values, inconsistent formats, duplicate records, outliers, or biased/imbalanced data. These issues can cause inaccurate predictions or unfair outcomes. For example, if your dataset has far more “approved” than “denied” loans, a model may appear accurate by predicting “approved” too often, while failing the minority cases that matter most.
Exam Tip: When a scenario mentions “inconsistent data,” “different units,” “missing values,” or “not enough representative examples,” the right direction is usually data preparation/cleaning—not choosing a fancier algorithm.
Feature engineering also includes encoding categories (e.g., product type), scaling numeric values, and extracting time-based signals (day of week, season) for forecasting. In AI-900 terms, you don’t need to know the math—just the reason: better features and cleaner data generally improve model performance and reliability.
Azure Machine Learning (Azure ML) is the primary Azure service for building, training, and deploying ML models. AI-900 expects you to recognize it as the “hub” for end-to-end ML work, especially when you need experiment tracking, model management, and deployment options.
The central container is the Azure Machine Learning workspace. Think of it as the project home where you manage assets and governance: experiments/runs, data connections, compute targets, models, endpoints, and monitoring integrations. If a question asks where you “organize and manage ML resources,” the workspace is usually the answer.
At a high level, Azure ML includes:
Exam Tip: If the scenario mentions “track experiments,” “reproduce runs,” “register a model,” or “deploy a model as a service,” it’s pointing to Azure Machine Learning rather than a prebuilt cognitive service.
AI-900 also tests workflow choices conceptually. Azure ML supports a visual drag-and-drop designer, a guided Automated ML experience, and a code-first approach (SDK/CLI) for teams needing full control. The exam does not require syntax—only when you would choose each.
Automated ML (AutoML) in Azure ML helps you rapidly build a model by trying multiple algorithms and hyperparameters, then selecting the best candidate based on a metric (accuracy, AUC, RMSE, etc.). For AI-900, the core concept is: you supply labeled data and define the task (classification, regression, forecasting), and AutoML automates much of the experimentation.
AutoML is a strong fit when you want a good baseline quickly, have limited ML expertise, or need to compare models efficiently. In contrast, the designer is ideal when you want a visual pipeline for data prep and training without writing code, while code-first is best when you need custom logic, advanced control, or integration into software engineering workflows.
Responsible ML concepts show up in questions about transparency, fairness, and explainability. Even at a non-technical level, you should know the purpose of interpretability: helping humans understand which inputs influenced predictions and why. This supports trust, debugging, and compliance in regulated decisions (credit, hiring, healthcare).
On AI-900, “responsible ML” is often tested as selecting actions that reduce harm: using representative datasets, monitoring performance drift, and being able to explain model decisions at a high level.
This section prepares you for the style of AI-900 questions without listing specific quiz items. Expect short scenarios where you must pick the ML type (classification, regression, clustering, forecasting) and then choose the right Azure ML concept (workspace, AutoML, designer, or code-first). The exam rewards fast recognition of keywords, but the safest method is to translate the scenario into “What is the output?” and “Do we have labels?”
For ML type selection, practice these mental checks: If the output is “yes/no” or a named category, it’s classification. If it’s a number, it’s regression—unless time is central, then it’s forecasting. If you are asked to “find segments” without outcomes provided, it’s clustering.
For Azure ML concept selection, look for cues: If the question emphasizes centralized management, governance, tracking runs, and deployments, it’s the workspace context. If it emphasizes trying multiple models automatically to find the best, it’s Automated ML. If it emphasizes visual steps and a no-code pipeline, it’s designer. If it emphasizes full control, custom training logic, or integration into development workflows, it’s code-first.
Use these checks to eliminate distractors quickly: any option that ignores data splitting/validation when asked about evaluation is likely wrong; any option that claims a model is “good” solely because training metrics are high is a red flag; and any option that treats clustering as a prediction of known labels is conceptually incorrect.
1. A retail company wants to predict the total revenue for each store next month based on historical sales, promotions, and local events. Which machine learning approach is most appropriate?
2. A bank wants to group customers into segments based on spending behavior and product usage, but it does not have predefined segment labels. Which type of machine learning should the bank use?
3. A team trains a model that performs extremely well on the training data but performs poorly on new, unseen data. In non-technical terms, what is the most likely issue, and what practice helps detect it?
4. A non-technical analyst wants to build and compare several models in Azure Machine Learning with minimal code, while automatically handling algorithm selection and hyperparameter tuning. Which workflow should they choose?
5. You are reviewing an Azure Machine Learning workspace at a high level. Which statement best describes what the workspace provides?
Computer vision questions on AI-900 are rarely about algorithms; they are about selecting the correct workload type and mapping it to the right Azure service family. Expect scenario-based prompts that describe a business need (for example, “find damaged items in warehouse photos” or “extract text from scanned invoices”) and ask what kind of computer vision task it is, and which Azure capability fits. This chapter teaches you to recognize the patterns quickly: classification vs. detection vs. segmentation vs. OCR; general image analysis vs. document reading; and when “custom” becomes necessary.
The exam also tests that you understand boundaries and responsible use. Vision solutions often handle biometrics or sensitive imagery, so you must be able to identify when consent, transparency, and restricted use apply—especially around facial analysis. Keep your focus on the objective: choose appropriate Azure AI Vision services for common vision workloads, and explain the difference between OCR/document reading and form-like extraction at a conceptual level.
Exam Tip: When you see the words “label the whole image,” think classification. When you see “locate items” or “draw boxes,” think object detection. When you see “pixel-level outline,” think segmentation. When you see “extract printed/handwritten text,” think OCR/Read.
Practice note for Recognize when to use image classification, object detection, and OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain Azure AI Vision capabilities and typical business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate OCR/document reading and form-like extraction at a conceptual level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand facial analysis considerations and responsible use expectations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Complete exam-style practice for 'Computer vision workloads on Azure': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize when to use image classification, object detection, and OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain Azure AI Vision capabilities and typical business scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate OCR/document reading and form-like extraction at a conceptual level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand facial analysis considerations and responsible use expectations: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 expects you to distinguish the major computer vision workload types based on the output required. In image classification, the model assigns one or more labels to the entire image (for example, “contains a cat” or “this is a product photo of shoes”). Classification is the right choice when location does not matter—only whether the image belongs to a category.
In object detection, the solution returns both a label and a location (typically a bounding box) for each detected object. Choose detection when the scenario includes counting items, locating hazards, finding inventory on shelves, or identifying where something appears in a frame (for example, “detect helmets on workers and mark their positions”).
Segmentation goes beyond boxes and returns a pixel-level mask—useful when you need precise outlines (for example, separating a background from a foreground object, or identifying the exact area of a defect on a surface). Segmentation is less common on AI-900 than classification/detection, but it appears as a conceptual option.
OCR (optical character recognition) is about extracting text from images: printed text, handwriting, signs in a photo, or text in scanned documents. On exam items, the clue is that the desired output is text, not “understanding of the image.”
Common trap: If a prompt says “identify all products in an image,” many learners jump to classification. But “all products” implies multiple objects and usually location/counting—object detection is the better match. Another trap is confusing OCR with “key-value extraction.” OCR returns text; structured field extraction is a document intelligence pattern (covered later in this chapter).
For AI-900, treat Azure AI Vision as the primary service family for general image understanding and text extraction from images. The exam often frames choices as “use a prebuilt capability” versus “train a custom model.” Azure AI Vision supports common tasks like tagging images, describing an image, detecting objects, and reading text (OCR). These capabilities are typically accessed through an endpoint and return JSON results (tags, captions, bounding boxes, or extracted text), which makes them easy to integrate into applications.
A critical concept is the difference between image analysis and Read/OCR. Image analysis focuses on visual content (what objects or scenes are present). Read/OCR focuses on transcribing text and often returns a hierarchy like pages, lines, and words with coordinates.
Exam Tip: If the scenario asks for “generate a caption,” “return tags,” or “detect common objects,” lean toward Azure AI Vision’s image analysis capabilities. If it asks to “extract text from photos or scanned pages,” select OCR/Read. Don’t overcomplicate with Azure Machine Learning unless the prompt says you need to build/train/manage your own model end-to-end.
Common trap: “Analyze a document” can mean two different things. If the goal is simply to read the text, Azure AI Vision Read fits. If the goal is to extract structured fields like invoice totals or line items, that’s typically a document extraction workload (conceptually different from plain OCR).
AI-900 expects you to know when prebuilt models are enough and when you should customize. Prebuilt vision models work well for broad, common concepts (everyday objects, generic scenes, common text). Customization becomes necessary when your categories are domain-specific, visually subtle, or not covered by general labels (for example, identifying a specific part number, a proprietary product variant, or distinguishing between similar defect types).
Customization also matters when your environment is unique: specialized camera angles, unusual lighting, medical or industrial imagery, or strict accuracy requirements. In these cases, you typically gather labeled examples and train a model for your specific classes (classification) or for locating your specific items (detection). On the exam, the signal words are “custom,” “company-specific,” “unique product,” “specialized,” or “not supported by built-in tags.”
Exam Tip: Look for the phrase “needs to recognize our own categories.” That is your cue that a custom vision approach is likely expected. Conversely, if the scenario is generic (tourism photos, standard retail items, common animals), prebuilt Vision is usually the intended answer.
Common trap: Learners assume “custom” automatically means “Azure Machine Learning.” For AI-900, customization might be described at a high level without requiring you to choose an ML platform. If the question is testing service selection for vision workloads, your best move is to choose the vision customization option when the requirement is domain-specific labels, not necessarily an end-to-end ML platform.
Text-in-image scenarios split into two conceptual levels that the exam loves to contrast. Level one is OCR/document reading: you want the words. This includes signs in photos, screenshots, scanned letters, or simple PDFs where returning the full text is enough for search, indexing, or accessibility.
Level two is form-like extraction: you want the meaning of the document’s structure—fields and values (for example, vendor name, invoice total, tax, dates) and sometimes repeating groups (like line items). Receipts, IDs, and invoices are classic prompts that hint you may need more than raw text. The exam does not require implementation details, but it does test whether you recognize that “extract total amount and merchant name from receipts” is not the same as “read the receipt text.”
Exam Tip: If the prompt includes “key-value pairs,” “fields,” “tables,” “line items,” or “populate a database,” interpret it as form/document extraction. If it says “make the document searchable” or “extract all text,” interpret it as OCR/Read.
Common trap: Choosing image classification for document problems. Documents are rarely “classified by pixels”; they are interpreted by text and layout. When the asset is a receipt/invoice form, the right mental model is text + structure.
Responsible AI is part of AI-900, and vision workloads are where it becomes concrete. You must be able to explain why facial analysis and biometric-like scenarios require extra care. Key concerns include fairness (performance differences across demographic groups), privacy (images are personal data), transparency (people should know when vision AI is in use), and security (protect stored images and extracted data).
Facial analysis scenarios may involve detecting a face, analyzing attributes, or verifying identity. Even if the exam question is not technical, it may ask what consideration is most important. In those cases, think: consent, purpose limitation, and avoiding sensitive inferences. If a solution uses faces for access control, it must include user consent, strong governance, and fallback mechanisms.
Exam Tip: When you see “identify employees,” “track customers,” “detect emotion,” or “infer sensitive traits,” pause—these are high-risk prompts. The best answer often emphasizes responsible use (consent, privacy safeguards, bias evaluation) over raw capability.
Common trap: Treating “can the service do it?” as the only decision factor. On AI-900, the right answer may highlight that a use case is sensitive or requires strict controls, even if the technology exists. Also watch for answers that recommend collecting more personal data than needed; “data minimization” is a safe guiding principle.
AI-900 vision questions are primarily service-selection and workload-identification items. Your strategy is to translate the business wording into the output type, then map that output to the simplest matching Azure capability. Start by underlining what the system must return: a label, a location, a pixel mask, or text/fields.
Use this exam-style decision flow: if the output is text, choose OCR/Read. If the output is labels for an entire image, choose classification. If you need locations or counts, choose object detection. If you need precise shapes, choose segmentation. Then ask: are the categories generic (prebuilt Vision) or domain-specific (custom vision approach)? Finally, apply responsible AI checks if people, faces, IDs, or surveillance-like monitoring appear.
Exam Tip: Eliminate distractors by asking, “Does this choice produce the required output?” Many wrong options are plausible services but do not return the needed structure (for example, recommending image tagging when the scenario needs bounding boxes, or recommending OCR when the scenario needs object counting).
Common trap: Over-selecting complex platforms. If a prebuilt Azure AI Vision capability satisfies the requirement, that is usually the exam’s intended answer. Reserve Azure Machine Learning-style answers for prompts that explicitly say you must build, train, and manage custom models with your own data pipeline.
As you practice, also look for hidden constraints: “in real time” hints at fast inference; “on-device” might imply edge deployment; “regulated documents” hints at governance and privacy controls. AI-900 will not require deep architecture, but it will reward choosing the most appropriate, least-assumptive service for the described scenario.
1. A retail company wants to automatically tag product photos as either "shoe", "shirt", or "hat" so that users can filter the catalog. The images contain only one main product per photo. Which computer vision workload best fits this requirement?
2. A warehouse team captures photos of pallets. They need to identify each box in the image and return its position so the system can count boxes and highlight missing ones. Which workload should you choose?
3. A company scans invoices and wants to extract printed and handwritten text lines from the images so that the text can be searched and archived. Which Azure capability is the most appropriate starting point?
4. A finance department receives many different vendor invoice layouts. They want to extract specific fields such as invoice number, total amount, and due date as structured key-value data, even when the text appears in different positions. Conceptually, which approach best matches this requirement?
5. A gym wants to deploy cameras that identify specific members as they enter, without requiring badges. Which consideration best aligns with responsible AI expectations and typical Azure restrictions for facial analysis scenarios?
This chapter targets the AI-900 objectives around Natural Language Processing (NLP) workloads and Generative AI workloads on Azure. On the exam, you are rarely asked to implement anything; instead, you are tested on recognizing the workload type from a business scenario and selecting the most appropriate Azure service family (Azure AI Language vs. Azure OpenAI, and when “classic NLP” is enough).
Expect scenario wording like “analyze customer feedback,” “extract company names,” “summarize call transcripts,” “build a chat experience,” or “generate marketing copy.” Your job is to map the business need to an NLP task (classification, extraction, summarization, translation) or to a generative task (content creation, Q&A with grounding), then identify the Azure product category that fits.
Exam Tip: When the requirement is to label or extract information from existing text (sentiment, key phrases, entities), think Azure AI Language. When the requirement is to create new text (draft, rewrite, brainstorm) or follow complex instructions conversationally, think Azure OpenAI—then add safety, grounding, and responsible use controls.
Practice note for Map business needs to NLP tasks: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Describe Azure AI Language capabilities and conversation/assistant use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts: prompts, tokens, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Describe Azure OpenAI use cases and responsible AI considerations for content generation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Complete mixed-domain practice for 'NLP workloads' and 'Generative AI workloads': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Map business needs to NLP tasks: sentiment, key phrases, entities, summarization: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Describe Azure AI Language capabilities and conversation/assistant use cases: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI concepts: prompts, tokens, grounding, and safety: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Describe Azure OpenAI use cases and responsible AI considerations for content generation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Complete mixed-domain practice for 'NLP workloads' and 'Generative AI workloads': document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
NLP workloads turn human language into structured signals or transformed text. AI-900 commonly tests four patterns: classification, extraction, summarization, and translation. Your first step in any scenario is to identify which of these the business wants.
Classification assigns a label to text. Typical labels include sentiment (positive/neutral/negative), topic categories, spam/not spam, or urgency levels. In business language, look for verbs like “categorize,” “route,” “detect tone,” or “flag.” Sentiment analysis is the classic example: analyzing product reviews to understand customer satisfaction.
Extraction pulls structured items out of text: key phrases, named entities (people, organizations, locations), or custom entities such as order numbers. Look for “extract,” “identify,” “pull out,” “tag names,” or “find mentioned products.” Key phrase extraction is not summarization—it returns short phrases, not a coherent paragraph.
Summarization reduces long text to a shorter representation, such as meeting notes, call center transcript summaries, or executive briefs. The exam may describe “too much text to read” or “provide a short overview.” Summarization can be extractive (selecting sentences) or abstractive (rewriting), but at fundamentals level you mainly need to recognize the use case.
Translation converts text from one language to another. Don’t confuse translation with transliteration (changing script) or with language detection (identifying the language). The scenario usually mentions “support multiple languages” or “translate support articles.”
Exam Tip: If the scenario’s output can be stored cleanly in columns (SentimentScore, EntityName, EntityType), that’s a strong signal for classic NLP analysis rather than generative AI.
Azure AI Language (part of Azure AI services) is the go-to choice on AI-900 for “analyze existing text” scenarios. It provides pre-built NLP capabilities that return structured results without you building a model from scratch. Typical exam scenarios include customer feedback analysis, document tagging, compliance monitoring, and insights extraction from emails, chats, or support tickets.
Map the business needs you see in prompts to these common Azure AI Language capabilities:
The exam also expects you to recognize when a pre-built analysis service is more appropriate than training custom ML. If the requirement is common and general (sentiment, key phrases, standard entities), choose Azure AI Language rather than Azure Machine Learning.
Common trap: Overreaching to generative AI. If the question asks to “extract key phrases and entities” from thousands of documents, using a generative model is usually unnecessary and costlier; structured NLP analysis fits better and is easier to validate at scale.
Exam Tip: Watch for wording like “return a score,” “identify entities,” “extract,” “detect,” or “classify.” Those verbs align to Azure AI Language analysis outputs and are strong hints for the correct service selection.
Conversational AI on AI-900 is assessed at a concept level: you should understand what a bot is, what intents and utterances are, and why conversation scenarios differ from one-off text analysis. A bot is an application that interacts with users via chat or voice to answer questions or complete tasks (checking order status, resetting passwords, scheduling appointments).
An utterance is what a user says or types (for example, “Where is my shipment?”). An intent is the goal behind that utterance (for example, TrackOrder). Many different utterances can map to the same intent. Conversations may also capture entities (slots) such as order number, date, city, or product name to complete the task.
On the exam, conversational scenarios often include requirements like “maintain context,” “handle follow-up questions,” “route to human agent,” or “provide consistent answers.” In classic bot design, you typically detect intent and extract entities, then call back-end systems. In generative experiences, you may instead rely on an LLM to interpret the user request—but you still need guardrails and grounding (covered later).
Exam Tip: If the scenario emphasizes transactional outcomes (book, order, reset) and extracting structured fields, think “intent + entities” design. If it emphasizes drafting explanations or flexible language generation, that leans toward generative AI.
Generative AI uses large language models (LLMs) to produce new content: text, summaries, answers, or transformations (rewrite, translate, classify via instructions). AI-900 expects you to know the core terms: prompts, tokens, temperature, grounding, and hallucinations.
A prompt is the input instruction and context you send to the model. Prompts can include role guidance (“You are a customer support agent”), constraints (“Respond in 3 bullet points”), and data (“Here is the product policy…”). Models process text in tokens (chunks of characters/words). Token limits matter because long prompts plus long responses can exceed the model’s context window—an exam-relevant concept when you see “very large documents” or “long chat history.”
Temperature controls randomness/creativity: lower values typically produce more deterministic, conservative outputs; higher values produce more varied outputs. On AI-900, don’t overthink numbers—just know the direction: lower for consistency (policy responses), higher for brainstorming (marketing ideas).
Hallucinations are plausible-sounding but incorrect outputs. They are not “bugs you can ignore”; they are a known risk of LLMs, especially when asked for facts beyond provided context. This is why grounding matters: connecting the model to trusted data (documents, databases) so responses are anchored in approved sources.
Common trap: Treating LLM output as guaranteed truth. If the scenario requires accuracy (regulatory, medical, financial), expect the best answer to include grounding and human review, not just “use an LLM to answer.”
Exam Tip: Look for keywords like “generate,” “draft,” “rewrite,” “create,” or “conversational responses.” Those are stronger indicators of generative AI than “extract” or “detect,” which usually indicate classic NLP analysis.
Azure OpenAI Service brings OpenAI models to Azure with enterprise features (identity, networking options, monitoring) and is positioned on AI-900 as the primary Azure service for generative AI workloads. Common capabilities you should recognize include generating text, summarizing, question answering, and assisting with content creation (emails, product descriptions) or conversational experiences.
A key concept is a deployment: in Azure OpenAI, you deploy a selected model and give it a deployment name your application calls. Exam items may describe “create a deployment,” “choose a model,” or “call the deployment endpoint.” You are not expected to code, but you should know that applications call deployments (not “the model” directly) and that different deployments can be tuned/configured for different scenarios.
Safe usage patterns are heavily emphasized in modern AI fundamentals. In scenario form, you may see: “prevent harmful content,” “avoid leaking sensitive data,” “ensure responses follow policy,” or “reduce hallucinations.” Strong answers usually include:
Common trap: Assuming generative AI is the best default for any text task. If the goal is consistent structured extraction at scale, Azure AI Language is often more appropriate and easier to validate. Use Azure OpenAI when you need flexible language generation, instruction following, or natural conversational responses.
Exam Tip: If a question mentions “responsible AI,” “harmful content,” “jailbreaks/prompt injection,” or “need to cite sources,” that’s a cue to discuss safety controls and grounding rather than only model selection.
AI-900 “mixed-domain” questions often blend NLP and generative AI in one scenario to test your ability to pick the simplest correct approach. Your decision framework should be: (1) identify the desired output type, (2) decide whether structured analysis is sufficient, (3) if generating content, apply grounding and safety requirements.
Use these recognition patterns when you read a scenario:
Responsible AI scenarios typically test whether you know what can go wrong and what mitigations fit. If users may enter personal data, the safest answer includes data handling controls (minimize sensitive inputs), access control, and monitoring. If the model might fabricate details, include grounding to trusted sources and a review process. If the risk is harmful or biased content, include content filters and policy-aligned prompting.
Common trap: Answering with a service name but ignoring the constraint. For example, “use Azure OpenAI to answer employee policy questions” is incomplete if the question also states “must be accurate and only from the handbook.” The better exam answer couples Azure OpenAI with grounding to the handbook and safety controls.
Exam Tip: When two answers seem plausible, choose the one that (a) matches the output type and (b) explicitly addresses risks (hallucination, harmful content, data leakage) with practical mitigations.
1. A retail company wants to automatically determine whether each customer review is positive, negative, or neutral and then trend the results over time. Which NLP task and Azure service family best fit this requirement?
2. A legal team has thousands of contract paragraphs and needs to extract organization names and locations mentioned in the text to populate a database. Which capability should you use?
3. A support center wants to provide agents with short summaries of long call transcripts to reduce after-call work. The transcripts are already available as text. What is the most appropriate approach?
4. A marketing team wants an assistant that can generate first-draft product descriptions from bullet points and rewrite them in different tones. Which Azure service family is the best fit?
5. A company is building a chatbot that answers employee questions using the company’s internal policy documents. They want responses to be based on those documents and to reduce the risk of the model making up facts. Which concept best addresses this requirement?
This chapter is your “dress rehearsal” for AI-900: you will simulate the real test, run two mixed-domain mock passes, diagnose weak spots, and finish with a tight review sprint. AI-900 rewards broad recognition more than deep engineering detail, so your goal is consistent service selection and clear mapping of scenarios to workloads: machine learning, computer vision, NLP, generative AI, and responsible AI. This chapter aligns directly to the exam outcomes: describing AI workloads and common solution scenarios; knowing when to use Azure Machine Learning; selecting Azure AI Vision and Azure AI Language capabilities; and understanding generative AI on Azure (Azure OpenAI) with responsible use considerations.
You will practice answering the way the exam expects: identify the workload first, then the Azure service family, then any key constraints (data type, real-time vs batch, custom vs prebuilt, governance, and safety). Many wrong answers on AI-900 are “near misses” from the right product family but wrong feature tier—your review clinic will train you to spot those distractor patterns.
Exam Tip: In AI-900, a correct answer often comes from one decisive keyword (e.g., “custom training,” “document extraction,” “speech,” “translation,” “chat completion,” “image tags,” “responsible AI”). Train yourself to circle that keyword mentally before you look at choices.
Use the sections below in order: simulate, attempt, attempt again, review deeply, remediate strategically, then lock in a calm exam-day routine.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Final review sprint and confidence calibration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Final review sprint and confidence calibration: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
To get value from a mock exam, you must simulate constraints, not just answer items casually. AI-900 is designed for breadth: you’ll shift rapidly among ML basics, Azure service selection, and responsible AI concepts. Your simulation rule: one sitting, timed, no notes, no searching. Use a quiet environment, a single screen, and a basic calculator only if your platform allows (most questions won’t require it).
Adopt a two-pass timing strategy. Pass 1: move quickly, answer what you know, and mark anything that needs rereading. Pass 2: return to marked items and force a decision using elimination. Do not spend disproportionate time on one confusing stem—AI-900 questions are intentionally short, and overthinking is a common failure mode.
Exam Tip: Your job is not to “prove” the answer; it’s to pick the best option given the scenario. If two options seem plausible, look for the constraint the exam writers want you to notice (custom vs prebuilt, structured vs unstructured data, online vs batch, or safety/governance requirements).
Simulate Microsoft-style question reading: first identify the workload category (prediction/classification/regression, vision, language, generative AI, responsible AI), then map to a service family (Azure Machine Learning, Azure AI Vision, Azure AI Language, Azure OpenAI). Only after that should you evaluate specific feature names. This prevents being pulled toward brand-recognition distractors.
Mock Exam Part 1 should feel like a realistic first half: it mixes “what is AI?” conceptual items with scenario sets that require service matching. Expect rapid transitions: one item may test supervised vs unsupervised learning, the next may ask which Azure service extracts text from images, and the next may check responsible AI principles.
When you see a business scenario, translate it into inputs and outputs. Example mental pattern: “We have labeled outcomes and want predictions” → supervised ML. “We want to group similar customers without labels” → clustering (unsupervised). “We want to detect anomalies in sensor streams” → anomaly detection workload; then decide whether the question wants the general concept or a specific Azure offering.
For vision and NLP scenarios, anchor on the data type: images/video vs text/conversation. Then decide prebuilt vs custom. Prebuilt services are often the best choice when the scenario says “quickly,” “no ML expertise,” or “minimal training data.” Custom training is more likely when the stem mentions “domain-specific,” “company-specific,” “custom labels,” or “train a model.”
Exam Tip: Scenario sets frequently hide the key in one noun: “documents,” “receipts,” “forms,” or “invoices” usually signals document analysis/extraction rather than general OCR. Likewise “sentiment” and “key phrases” are classic NLP tasks; “tags” and “object detection” are classic vision tasks.
As you complete Part 1, track your confidence per item (high/medium/low). Confidence tracking is not fluff—it powers your weak spot analysis later so you can distinguish true gaps from careless reading.
Mock Exam Part 2 should increase service-selection density: you will be asked to choose between closely related Azure options. This is where many non-technical candidates lose points—not because they don’t understand the workload, but because they confuse product families or choose a tool that is “capable” rather than “best fit.” Your discipline: pick the service that is purpose-built for the task described.
Machine learning selection: if the scenario emphasizes building, training, and managing models (datasets, experiments, pipelines, ML lifecycle), Azure Machine Learning is the safe anchor. If the scenario is simply “use AI without building models,” the exam often expects Azure AI services (Vision/Language) instead of AML. For generative AI, when the task is chat, summarization, or content generation with large language models, Azure OpenAI is the typical match.
Language scenarios often split into analysis vs conversation. “Extract entities, key phrases, sentiment” signals Azure AI Language. “Real-time speech-to-text or text-to-speech” is a different workload (speech), so do not let “language” wording trick you into choosing text analytics features when the input is audio. Vision scenarios similarly split: image understanding (tags, detection), text in images (OCR), and document extraction. Use the stem’s output requirement to decide.
Exam Tip: If the options include a general platform (like Azure Machine Learning) and a specialized cognitive service (Vision/Language/OpenAI), default to the specialized service unless the question explicitly requires custom model training, MLOps, or end-to-end ML lifecycle control.
Generative AI items often include responsible use constraints: you may need content filtering, prompt safety, or human oversight. Treat those as first-class requirements, not afterthoughts—AI-900 tests responsible AI basics as part of solution selection.
Your review clinic is where you gain points fastest. Do not just mark answers right/wrong; write a one-sentence rationale tied to an exam objective: “This is computer vision (image input), so choose Azure AI Vision for image analysis,” or “This is supervised learning with labels, so it’s classification/regression.” If you cannot explain it in one sentence, you likely don’t own the concept yet.
Learn distractor patterns. Pattern 1: “Wrong family, right buzzword.” Example: choosing Azure Machine Learning for simple sentiment analysis because it sounds advanced. Pattern 2: “Right family, wrong feature.” Example: selecting generic OCR when the scenario wants structured field extraction from invoices. Pattern 3: “Overfitting to one keyword.” Example: seeing the word “language” and picking text analytics even though the input is speech audio. Pattern 4: “Ignoring constraints.” Example: choosing a custom model approach when the stem emphasizes minimal data or rapid deployment.
Exam Tip: If two answers look correct, ask: which one requires fewer assumptions? AI-900 favors the most direct, managed, purpose-built solution that matches the stated input/output and constraints.
Also review responsible AI traps. The exam may test fairness, reliability, privacy/security, inclusiveness, transparency, and accountability. A common trap is treating responsible AI as optional guidance rather than design requirements. If the stem mentions sensitive domains (HR, lending, healthcare), expect the correct answer to emphasize governance, explainability, privacy, and human oversight.
Finally, watch for “scope traps”: AI-900 is fundamentals. If an answer choice dives into deep infrastructure or niche engineering, it’s often a distractor unless the stem explicitly calls for it.
After both mock parts, score by domain, not just total. Create a simple grid: ML fundamentals, Azure Machine Learning use cases, computer vision, NLP, generative AI/Azure OpenAI, and responsible AI. For each missed item, label the failure type: (A) concept gap, (B) service confusion, (C) careless reading, or (D) overthinking. Your remediation plan depends on the type.
For concept gaps (A), revisit the workload definition and a single canonical example. For service confusion (B), build a “decision cue” list: what keyword pushes you toward Vision vs Language vs AML vs OpenAI. For careless reading (C), slow down on stems with negatives (e.g., “NOT,” “least likely”) and on questions that list multiple requirements. For overthinking (D), practice committing to the simplest managed service that meets requirements.
Exam Tip: Don’t remediate by rereading everything. Remediate by rewriting your own rules. Example: “If it’s unstructured text analysis → Azure AI Language; if it’s labeled prediction → Azure Machine Learning; if it’s image understanding → Azure AI Vision; if it’s content generation/summarization → Azure OpenAI.”
Plan a retake strategy even if you hope not to need it. If your mock score is borderline, schedule a short retake window (7–14 days) and focus on your lowest two domains first. Retaking without changing your study method usually reproduces the same mistakes; retaking after building decision rules and reviewing distractor patterns typically yields a quick improvement.
End this section by doing a “confidence calibration”: compare what you felt confident about versus what you actually got right. Your goal is to reduce false confidence (dangerous) and increase accurate confidence (stabilizing under time pressure).
On exam day, eliminate avoidable stress. Confirm your ID requirements and name matching exactly. If testing online, validate your environment early: stable internet, power, quiet room, cleared desk, and allowed peripherals only. If testing in a center, arrive early enough to complete check-in without rushing your mindset.
Pacing: commit to the same two-pass method you used in the mocks. Pass 1 builds momentum and secures easy points; Pass 2 is where you use elimination and constraints. If you catch yourself rereading the same sentence repeatedly, mark it and move—this is a known sign of overinvestment.
Exam Tip: For every question, force the “final mental model” in five seconds: (1) What workload is this? (2) What data type is used? (3) Do we need custom training or prebuilt? (4) Which Azure service family fits? (5) Any responsible AI or governance constraint?
Use a final review sprint the day before: refresh your decision cues, revisit your most-missed domain, and rehearse responsible AI principles with one real-world example each. Avoid last-minute deep dives into advanced engineering topics; AI-900 rewards clear fundamentals and service recognition.
Finish with confidence calibration: you don’t need perfection—you need consistency. If you can reliably map scenarios to workloads and choose the most direct Azure AI service while respecting responsible AI basics, you are operating at the level this exam is designed to certify.
1. A company wants a chatbot on its website that can answer questions using the company’s policy documents and must reduce the chance of unsafe or inappropriate responses. Which Azure service and feature set best fits this requirement?
2. You are reviewing a mock exam result and notice you frequently miss questions that involve extracting structured fields (invoice number, totals, vendor name) from scanned invoices. Which Azure AI capability should you choose in these scenarios?
3. A retail chain wants to forecast weekly demand for each store. They have historical sales data in a table and want a guided approach that can automatically try multiple algorithms and select the best model. Which service should they use?
4. A global support team wants to translate customer emails from multiple languages into English in near real time. Which Azure service capability best matches this need?
5. On exam day, you see a question describing an app that must detect objects (e.g., 'person', 'forklift') in images from a warehouse camera feed. The solution should use a prebuilt model, not custom training. Which option should you pick?