AI Certification Exam Prep — Beginner
A clear, non-technical path to passing Microsoft AI-900 on the first try.
This course is a beginner-friendly blueprint for passing Microsoft’s AI-900 (Azure AI Fundamentals) certification exam. It is designed for non-technical professionals—sales, operations, project managers, analysts, and anyone who needs to understand AI concepts and Azure AI services without becoming a developer. You’ll learn the language of AI, how to recognize the right workload for a business problem, and how Microsoft expects you to answer scenario-based questions.
The curriculum maps directly to the official AI-900 exam domains so you spend time only where it matters. You’ll build competence across:
The course is organized like a focused exam-prep book. Chapter 1 helps you start correctly: how to register, what the question formats look like, how scoring works, and how to plan your study time. Chapters 2–5 dive into the exam objectives with plain-language explanations and the kind of decision-making practice you need for AI-900 (for example, choosing the best workload type from a scenario). Chapter 6 finishes with a full mock exam experience plus a final review checklist and exam-day readiness plan.
AI-900 rewards clarity and correct selection—knowing what a workload is, what it’s used for, and what success looks like. This course emphasizes exam-relevant distinctions (classification vs regression, OCR vs image classification, classic NLP vs generative AI) and reinforces them through exam-style practice sets. You’ll also learn a repeatable method for reading scenarios, identifying keywords, and eliminating distractors—critical skills for Microsoft fundamentals exams.
If you’re new to certifications, begin by setting up your learning routine and account access. You can Register free to start your plan today, or browse all courses to compare learning paths. Then follow the chapters in order, complete the practice sets, and take the mock exam under timed conditions to confirm readiness.
This course is ideal for learners with basic IT literacy who want an accessible, structured path to Microsoft AI-900. No prior Azure experience is required—only a willingness to practice and review mistakes until the exam patterns feel familiar.
Microsoft Certified Trainer (MCT)
Jordan Whitaker is a Microsoft Certified Trainer who helps beginners pass Microsoft fundamentals exams through practical, exam-aligned instruction. He has coached professionals across business, operations, and customer-facing roles to understand Azure AI concepts and confidently sit AI-900.
AI-900 (Microsoft Azure AI Fundamentals) is designed to verify that you can speak confidently about common AI workloads and Azure’s AI service families—even if you are not a developer. This chapter orients you to what the exam measures, how the exam works, and how to build a practical study plan in 2–4 weeks. As an exam coach, my goal is to help you avoid “study noise” (too much depth in the wrong areas) and focus on what Microsoft actually tests: selecting the right AI approach, understanding core terminology, and recognizing responsible AI considerations in real scenarios.
You’ll also learn how to approach Microsoft-style questions. These questions often look like short business cases: a team wants to extract text from invoices, analyze customer sentiment, detect objects in images, or create a chatbot. Your job is to match the need to the correct AI workload and the right Azure capability (for example, OCR vs image classification; text analytics vs conversational AI; classical ML vs generative AI). The rest of the course will build the knowledge, but your results will improve immediately if you treat AI-900 as a pattern-recognition exam: identify the workload, identify constraints (latency, privacy, cost, accuracy), and choose the best-fit option.
Exam Tip: For AI-900, aim for breadth over depth. You do not need to design neural networks or write training code, but you do need to accurately classify scenarios and explain why an approach fits (or doesn’t).
Practice note for Understand the AI-900 exam format, timing, and scoring: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set up your test environment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a realistic 2–4 week study plan for beginners: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for How to approach Microsoft exam questions (strategy and pitfalls): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand the AI-900 exam format, timing, and scoring: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register for the exam and set up your test environment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Build a realistic 2–4 week study plan for beginners: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for How to approach Microsoft exam questions (strategy and pitfalls): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand the AI-900 exam format, timing, and scoring: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900 measures foundational understanding across five outcome areas you’ll see throughout this course: describing AI workloads and responsible AI; understanding fundamental machine learning concepts on Azure; identifying computer vision workloads; identifying natural language processing (NLP) workloads; and describing generative AI workloads (including Azure OpenAI and prompt concepts). If you are a non-technical professional—product manager, business analyst, sales engineer, program lead, or educator—AI-900 validates that you can collaborate with technical teams and make sound AI decisions without hand-waving.
What the exam is really testing is your ability to translate a business need into an AI workload category. You should be able to tell when a rules-based solution is sufficient (for example, deterministic validation logic) versus when AI adds value (for example, language understanding, image recognition, anomaly detection). You should also be able to recognize responsible AI concerns—fairness, reliability, privacy, transparency, and accountability—and how they affect solution selection.
Common Trap: Confusing “machine learning” as the answer for every scenario. Some questions want a simpler decision: rules-based logic, search, or a prebuilt AI service (for example, OCR) rather than custom model training. Your best answer is the one that meets requirements with the least complexity.
Registering correctly prevents last-minute problems that have nothing to do with your knowledge. AI-900 is scheduled through Microsoft’s certification dashboard and delivered via an exam provider (often Pearson VUE). You typically choose between an online proctored exam (taken at home/office) and a test center. Your choice should depend on stability and risk: online is convenient but less forgiving of environmental issues; test centers reduce technical variables but require travel.
Before you book, confirm the exam language, time zone, and name on your profile. A mismatch between your registration name and your ID is one of the most preventable reasons for being turned away. Plan for a valid, government-issued ID that meets the provider’s requirements (often photo ID; sometimes a second ID depending on region).
Exam Tip: Treat your exam environment like a dependency. If your environment fails, your score doesn’t matter. Do a full check: webcam, microphone, room lighting, and system test (if offered) at least 24 hours in advance.
Common Trap: Scheduling an online exam during a high-interruption time. Unexpected pop-ups, notifications, or people entering the room can trigger a proctor warning. Choose a time slot when you can fully control your space.
AI-900 uses Microsoft’s typical question styles: multiple choice, multiple response (“choose all that apply”), scenario-based items, and occasionally drag-and-drop or matching formats. The exam is designed to test understanding, not memorization of obscure facts, but you must be precise with terminology (for example, the difference between classification vs regression; OCR vs image analysis; entity recognition vs key phrase extraction; generative AI vs traditional NLP).
Microsoft does not always disclose exact scoring mechanics per item, and some exams may include unscored items for quality evaluation. Your pass strategy should assume that every question counts and focus on consistency. Read carefully for requirement words like “best,” “most cost-effective,” “least administrative effort,” “requires training data,” or “near real-time.” Those words are the scoring hinge.
Exam Tip: When two answers both look plausible, choose the one that requires the least custom work. AI-900 often rewards “use a prebuilt service” over “build a custom model” unless the prompt explicitly requires custom training or specialized outputs.
Common Trap: Over-reading the scenario and inventing requirements. Only solve the problem stated. If the question doesn’t mention multilingual needs, don’t pick an option solely because it supports multiple languages.
A realistic beginner plan for AI-900 is 2–4 weeks, depending on how many hours you can commit. The exam domains map cleanly to weekly themes. Your objective is not to become an engineer; it is to develop a “workload selection reflex” and a responsible AI lens. Build repetition by cycling: learn concepts, review terminology, then apply to mini-scenarios (without turning your prep into heavy labs).
2-week plan (accelerated, ~45–60 minutes/day): Week 1 covers AI workloads, responsible AI, and ML fundamentals (training vs inference, features/labels, evaluation metrics at a high level). Week 2 covers vision, NLP/speech, and generative AI concepts and use cases, with daily review of scenario mapping.
4-week plan (steady, ~30–45 minutes/day):
Exam Tip: Each week, create a one-page “decision map” (workload → typical tasks → common Azure service family). The exam rewards quick categorization more than deep implementation knowledge.
Common Trap: Spending too long on model math. AI-900 may reference metrics conceptually, but you rarely need formulas. Prioritize understanding what a metric implies (for example, when accuracy is misleading) over calculation.
Microsoft Learn is your primary study resource for AI-900 because it matches Microsoft’s vocabulary and scenario framing. Use it strategically: skim for structure, then re-read sections that map to exam outcomes (workloads, service selection, responsible AI, and generative AI). Your goal is to adopt Microsoft’s terms so you can recognize them in questions.
You may also explore the Azure portal to become familiar with how services are presented, but do so safely. Many Azure services can be used within free tiers or limited trials, yet it’s still possible to create billable resources accidentally. Build “no-cost habits” now: always check pricing pages, watch for paid SKUs, and delete resources immediately after experimentation.
Exam Tip: The portal helps you recognize service names and categories, which is valuable for multiple-choice elimination. However, AI-900 does not require you to memorize portal clicks. Focus on what the service does and when to use it.
Common Trap: Confusing “Azure Machine Learning” (a platform for building ML solutions) with “Azure AI services” (prebuilt APIs for vision/language/speech). On the exam, the correct choice often hinges on whether you need custom training and model management versus consuming a prebuilt capability.
Exam-day success is mostly discipline: eliminate preventable friction, manage your pace, and avoid mental errors. If taking the exam online, start early to complete check-in steps and room validation. If at a test center, plan arrival time with a buffer for traffic and check-in. Either way, ensure you are rested; AI-900 is not conceptually hard, but it is detail-sensitive.
Time management is about avoiding stalls. Don’t spend too long on one question because the exam is designed to test breadth. Use a two-pass method: answer what you know quickly, mark items you’re uncertain about (if the interface allows), and return later. When you revisit, apply elimination: remove answers that solve a different workload, require unnecessary complexity, or ignore the scenario constraints.
Exam Tip: Watch for “keyword traps.” Words like “predict,” “forecast,” and “estimate” often indicate regression; “categorize” indicates classification; “extract text” points to OCR; “chat” indicates conversational AI; “generate” and “summarize” may indicate generative AI. Use these cues to orient fast.
Common Trap: Second-guessing correct answers after noticing unfamiliar terms. If your selected answer matches the workload and constraints cleanly, keep it unless you can articulate a stronger fit. Confidence comes from a repeatable method, not from recognizing every term.
1. You are creating a 3-week study plan for a beginner taking AI-900. Which approach best aligns with Microsoft’s expected level for this exam?
2. A practice question states: “A retail company wants to extract printed text from scanned invoices to populate a database.” Which workload should you identify first to choose the correct Azure capability on AI-900?
3. You are answering a Microsoft-style case question. The prompt describes requirements such as privacy constraints, latency expectations, and acceptable accuracy. What is the best first step to avoid common pitfalls on AI-900?
4. A colleague says: “To pass AI-900, I need to learn how to write model training code and tune hyperparameters in Python.” What is the most accurate response based on the AI-900 exam orientation?
5. A team is planning to take AI-900 remotely. They want to minimize test-day issues. Which action best fits recommended exam setup guidance covered in an orientation chapter?
This chapter maps directly to the AI-900 “Describe AI workloads” domain. The exam expects you to recognize common AI scenarios in plain language, pick the right workload type (vision, NLP, prediction, anomaly detection, generative), and explain why an AI approach is appropriate (or not) compared to rule-based automation. You are not being tested as a data scientist, but you are being tested on accurate workload selection, core terminology (AI vs machine learning vs deep learning), and responsible AI principles applied to realistic business situations.
As you study, keep one habit: translate every scenario into (1) the input data type (text, image, audio, tabular), (2) the desired output (label, number, extracted text, generated content, “is this normal?”), and (3) whether the solution needs learning from examples or can be expressed as deterministic rules. This chapter’s sections align to those decision points and to how AI-900 questions are typically written—short scenario prompts with one or two keywords that reveal the workload.
Practice note for Recognize common AI workloads and where they fit in business: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate AI, machine learning, and deep learning in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Describe AI workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize common AI workloads and where they fit in business: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate AI, machine learning, and deep learning in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Apply responsible AI principles to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Describe AI workloads (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize common AI workloads and where they fit in business: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Differentiate AI, machine learning, and deep learning in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
The AI-900 objective “Describe AI workloads” is about matching business problems to the right AI capability. The exam frequently uses everyday workplace examples—support tickets, invoice processing, product recommendations, defect detection, forecasting demand, summarizing meetings—and asks what kind of AI is being described. Your job is to name the workload and often the Azure approach (prebuilt service vs custom model) at a high level.
In plain language: AI is the umbrella term for systems that perform tasks requiring human-like intelligence. Machine learning (ML) is a subset where the system learns patterns from data instead of being explicitly programmed with rules. Deep learning is a subset of ML that uses multi-layer neural networks and is especially common for images, speech, and large language models. The exam trap is thinking these are competing choices; in reality, deep learning is one technique used within ML, and ML is one approach within AI.
Exam Tip: When a question mentions “trained on historical examples,” “model,” “features,” “evaluation,” or “improves with data,” that’s ML. When it mentions “neural network,” “transformer,” “embeddings,” “computer vision,” “speech recognition,” or “large language model,” that often implies deep learning—still ML, still AI.
Another common test angle is capability vs workload. “Optical character recognition” is a capability; the workload is typically “document processing” or “text extraction from images.” Focus on the task being performed and the input/output, not buzzwords.
Most AI-900 scenarios collapse into a few repeatable workload patterns. Learning these patterns makes multiple-choice questions much easier because distractors often swap similar terms (classification vs prediction, detection vs recognition). Start by identifying the output type.
Exam Tip: “OCR” is not classification; OCR is extraction of text from images. A common trap is choosing “text analytics” when the input is a scanned document. If the input is an image/PDF scan and the task is reading text, it’s a vision/document intelligence/OCR workload first, then NLP can be applied after the text is extracted.
Also watch for “recommendations.” Many learners assume recommendations are always generative AI. In exam framing, recommendations are typically prediction/ranking (ML) rather than generation—unless the system is asked to write a personalized message or explanation, which leans generative.
A high-value AI-900 skill is knowing when not to use AI. Rule-based automation (traditional programming, if/then logic, deterministic workflows) is often cheaper, easier to audit, and more reliable when the business logic is stable and can be expressed explicitly. AI becomes valuable when rules are hard to define, patterns are complex, or the environment changes.
Use a simple decision test the exam loves: if you can write a clear rule list that covers edge cases (“if X then Y”) and the inputs are structured and consistent, rules may be best. If you need to learn from examples (emails vary wildly, images differ by lighting, language is ambiguous), AI/ML is more appropriate.
Exam Tip: Questions often include the phrase “cannot be easily defined with explicit rules” as a clue to choose AI. Conversely, when the scenario says “business rules are well-defined” or “must be fully explainable and deterministic,” consider rule-based automation.
Common trap: assuming AI always improves accuracy. AI introduces model drift, bias risk, and probabilistic outputs. For compliance-heavy decisions (loan approvals, hiring), the exam expects you to think about human oversight, transparency, and responsible AI—even if an AI approach is technically possible.
AI-900 regularly tests responsible AI principles using short workplace scenarios. You’re expected to know the principles and match them to the risk described. Microsoft commonly frames these as: fairness, reliability & safety, privacy & security, transparency, and accountability.
Exam Tip: When a scenario includes “customers must be informed,” that’s transparency. When it includes “who is responsible if it fails,” that’s accountability. When it mentions “personal data,” jump to privacy & security. Many wrong answers swap transparency and accountability—anchor on “disclosure/explanation” (transparency) vs “ownership/oversight” (accountability).
Practical application matters: responsible AI is not an afterthought. It influences data collection (representative datasets), evaluation (performance across groups), deployment (monitoring drift), and user experience (clear disclosures and feedback loops). Even as a non-technical professional, the exam expects you to recognize that risk controls are part of the workload design.
AI-900 does not require you to implement solutions, but it does test whether you can choose between prebuilt AI services and custom ML. The key tradeoff is speed and simplicity vs specialization and control.
Prebuilt services (Azure AI services) are ideal when the task is common and you want fast time-to-value: image tagging, OCR, face blurring, speech-to-text, translation, sentiment analysis, key phrase extraction, and basic chatbots. These are typically consumed via simple APIs and require minimal ML expertise.
Custom ML (Azure Machine Learning) is best when your data is unique, labels are specific to your business, or you need higher accuracy for a narrow domain (for example, classifying internal ticket types, predicting equipment failure based on proprietary sensor patterns). Custom also supports more control over training, evaluation, and deployment.
Generative AI workloads often use foundation models via Azure OpenAI (for summarization, drafting, Q&A over enterprise content, copilots). You may still combine this with custom components like retrieval (search over your documents) or safety filters. The exam frequently tests that “prompting a foundation model” is not the same as “training a custom ML model.”
Exam Tip: If the prompt says “no labeled data available” and the task is standard (OCR, translation, speech recognition), prebuilt is usually correct. If the prompt emphasizes “company-specific categories,” “custom labels,” or “trained on our historical records,” expect Azure Machine Learning/custom training. If it emphasizes “summarize,” “draft,” “generate,” “copilot,” expect Azure OpenAI/foundation model usage.
AI-900 questions reward disciplined reading. Many candidates miss points by picking the first familiar term (like “AI”) instead of the best-fitting workload. Use a two-pass method: (1) identify the input modality, (2) identify the output/goal, then map to a workload pattern.
Exam Tip: Watch for multi-step scenarios. Example structure: “extract text from invoices and then classify the request type.” The correct approach is often a pipeline: OCR (vision/document intelligence) first, then NLP/classification. The exam may ask for the best first step or primary workload—answer the step the question actually asks for.
Common traps include confusing “translation” (NLP prebuilt) with “summarization” (generative), confusing “object detection” (locating items in images) with “image classification” (one label for the whole image), and assuming all chat experiences require generative AI (many are intent-based bots). Your advantage as a non-technical pro is to focus on business intent and choose the simplest tool that satisfies it, while recognizing where responsible AI must be addressed.
1. A retail company wants to automatically read product serial numbers from photos taken by warehouse staff using mobile phones. Which AI workload should you use?
2. A bank has years of historical transaction records labeled as fraudulent or legitimate. The bank wants a system that classifies new transactions into one of those two categories. Which approach best describes the solution?
3. A manufacturing plant streams temperature and vibration sensor readings from equipment. The goal is to flag unusual patterns that may indicate impending failure, even when there are few examples of past failures. Which AI workload is the best fit?
4. You are describing AI concepts to a business stakeholder. Which statement correctly differentiates AI, machine learning, and deep learning in plain language?
5. A healthcare provider deploys an AI model to help prioritize patient follow-up calls. During review, staff notice the model performs worse for a specific demographic group. Which Responsible AI principle should be addressed first?
This chapter maps to the AI-900 objective: explain fundamental principles of machine learning (ML) on Azure, including how models are trained, evaluated, and operationalized with core Azure Machine Learning concepts. The exam targets your ability to recognize ML workloads, use the right learning approach, and select appropriate evaluation metrics. You are not expected to code, but you are expected to interpret scenarios and choose the best option from Azure-flavored terminology (dataset, compute, pipeline, endpoint).
A reliable way to succeed on AI-900 is to read each question as a “business need + data available + success definition” problem. That pattern will tell you whether you need classification or regression, which metric matters, and whether Azure Machine Learning (Azure ML) is the right tool versus a rules-based approach.
We’ll cover supervised/unsupervised/reinforcement learning, explain training/validation/testing and overfitting in simple terms, choose the right metrics for classification vs. regression, then connect the ideas to Azure ML basics. You’ll finish with a practice set and a mini-case (delivered in your course assets) that trains the exact exam skill: mapping a business need to an ML approach.
Practice note for Understand supervised, unsupervised, and reinforcement learning: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain training, validation, testing, and overfitting in simple terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose the right metrics for classification vs regression: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML fundamentals on Azure (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-case: map a business need to an ML approach: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand supervised, unsupervised, and reinforcement learning: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain training, validation, testing, and overfitting in simple terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Choose the right metrics for classification vs regression: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: ML fundamentals on Azure (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mini-case: map a business need to an ML approach: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
On AI-900, “fundamental principles of ML on Azure” means you can explain what a model is, how it learns from data, how you check if it generalizes, and which Azure components help manage that lifecycle. The exam does not expect algorithm math; it expects correct identification of the workload and correct evaluation/implementation choices.
Most scenario questions hint at ML when the solution must learn patterns from historical examples (for example, predicting churn, classifying emails, forecasting sales). In contrast, if the decision can be expressed as fixed logic (for example, “if customer age < 18, deny loan”), rules may be better. When the data changes or the patterns are complex, ML often wins because it adapts through retraining.
Exam Tip: Watch for language that implies learning from examples (“based on past outcomes,” “predict,” “classify,” “forecast,” “probability”). That’s your signal to think ML rather than rules.
The chapter lessons connect directly to test objectives: (1) define supervised/unsupervised/reinforcement learning, (2) explain training/validation/testing and overfitting, (3) choose metrics for classification vs regression, and (4) recognize Azure ML building blocks. Your practice set should reinforce interpreting prompts and discarding distractors that use the wrong learning type or wrong metric.
AI-900 frequently tests vocabulary. A feature is an input variable used to make a prediction (for example, customer tenure, monthly spend). A label is the known correct output in your historical training data (for example, “churned: yes/no” or “next month revenue: 1250”). A model learns a mapping from features to labels during training.
Inference is the act of using a trained model to generate a prediction on new data. Many exam questions disguise inference as “scoring,” “predicting,” or “running the model in production.” If the scenario says “new customers arrive daily and you must assign a risk score,” that’s inference.
The model lifecycle (in simple exam terms) is: collect/prepare data → train → evaluate → deploy → monitor → retrain. Azure ML exists to make that lifecycle manageable with tracked experiments, repeatable pipelines, and deployable endpoints.
Exam Tip: If the question mentions “known outcomes” or “historical labeled data,” you’re in supervised learning territory. If it mentions “no labels available” but wants grouping or structure discovery, you’re in unsupervised learning territory.
Common traps: mixing up features vs labels (labels are not inputs), and confusing training with inference (training uses many examples to learn; inference uses the trained model to predict on new records).
Supervised learning uses labeled data (features + known label) to learn to predict the label for new cases. On AI-900, supervised learning splits into two high-frequency workloads: classification (predict a category like fraud/not fraud) and regression (predict a number like price or demand). The exam will often give you the target column; if it’s categorical, think classification; if it’s numeric and continuous, think regression.
Unsupervised learning uses data without labels to discover structure. The most common concept tested is clustering, where the model groups items based on similarity (for example, segmenting customers into groups with similar purchasing behavior). Unsupervised learning is not “less accurate”; it’s solving a different type of problem where no ground-truth labels exist.
Reinforcement learning (RL) is learning by trial and feedback—an agent takes actions and receives rewards or penalties. AI-900 treats RL as a conceptual category (you should recognize it), often described with phrases like “maximize reward,” “learn a policy,” “game/robot navigation,” or “dynamic decision-making.” It’s less about Azure-specific tooling and more about identifying that the learning happens through interaction rather than labeled examples.
Exam Tip: If the scenario includes “recommend the next best action” and mentions rewards over time (not just predicting a label), consider reinforcement learning. If it’s just choosing among products based on past clicks, that may still be supervised learning (predict click/no click) or ranking, not RL.
In your mini-case, practice translating business needs into one of these three learning types before thinking about Azure services. This avoids a classic mistake: jumping to “use clustering” when the problem actually provides labeled outcomes, or choosing classification when the output is a numeric forecast.
AI-900 expects you to match evaluation metrics to the ML task and to interpret “what good looks like” in business terms. For classification, you’ll see the confusion matrix: true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN). These numbers drive metrics like accuracy, precision, and recall.
Accuracy is the percentage of correct predictions overall. It can be misleading when classes are imbalanced (for example, fraud is rare). Precision answers: “Of the items predicted positive, how many were actually positive?”—useful when false positives are costly (flagging legitimate transactions). Recall answers: “Of the actual positives, how many did we catch?”—useful when missing a positive is costly (missing a fraudulent transaction or a dangerous defect).
Exam Tip: When the scenario emphasizes “don’t miss any real cases,” prioritize recall. When it emphasizes “don’t overwhelm staff with false alarms,” prioritize precision.
For regression, accuracy/precision/recall don’t apply. Instead, you’ll see error-based metrics such as MAE (mean absolute error) and RMSE (root mean squared error). MAE is the average absolute difference between predicted and actual values, and it’s easier to explain in business terms (“on average, we’re off by $12”). RMSE penalizes larger errors more heavily, making it sensitive to outliers.
Training/validation/testing connects directly to evaluation. The training set teaches the model; the validation set helps tune choices (like model parameters); the test set provides an unbiased final check. Overfitting means the model performs well on training data but poorly on new data—usually because it learned noise rather than general patterns.
Azure Machine Learning (Azure ML) is the Azure service most associated with building and managing ML models end-to-end. AI-900 won’t test deep configuration, but it does test recognition of core pieces and what they’re for.
A dataset (or data asset) is how you register and manage training data references. Compute refers to the resources used to run training or inference (compute instances/clusters). Questions may describe “scaling training” or “running experiments” and expect you to choose compute rather than storage or deployment.
Pipelines are repeatable workflows that connect steps like data prep, training, and evaluation. If a scenario requires repeatability, auditing, or scheduled retraining, pipelines are a strong fit. Deployment is often described as hosting the model behind an endpoint for real-time or batch predictions. While AI-900 stays high-level, remember the big idea: training happens in an experiment; inference happens at a deployed endpoint.
Exam Tip: If the question says “make predictions for new data in an app,” think “deploy a model as an endpoint.” If it says “track runs and compare models,” think “experiments” and “registered models,” not deployment.
This is where your practice set helps: many distractors use Azure services that sound plausible but don’t match the ML lifecycle step. Your job is to map the requirement to the stage: data → train → evaluate → deploy → monitor.
Two of the most common AI-900 mistakes are (1) choosing a metric that doesn’t match the task, and (2) describing a process that accidentally leaks information from the future into training.
Metric mismatch: Using classification metrics (accuracy/precision/recall) for regression problems, or using MAE/RMSE for classification. The exam frequently hides the task type inside business wording. If the output is a number (price, demand, duration), it’s regression—use MAE/RMSE. If the output is a category (approve/deny, spam/not spam), it’s classification—use confusion-matrix metrics. If the output is grouping with no labels, it’s clustering—evaluation is different and usually described qualitatively or with clustering-specific measures (not a core AI-900 focus).
Exam Tip: Before reading answer options, identify the target output type (category vs number vs no label). This one step eliminates most wrong choices.
Data leakage occurs when information that wouldn’t be available at prediction time is used during training (for example, using “cancellation date” to predict churn, or using post-outcome fields to predict the outcome). Leakage inflates validation/test scores and leads to failure in production. The exam may describe overly “perfect” results or include a feature that is clearly derived from the label.
Another leakage variant is improper splitting: if you randomly split time-series data without respecting chronology, the model can “peek” at future patterns. For non-technical pros, remember the rule: training data must reflect what you would know at the moment you need to make the prediction.
Use your mini-case to practice a leakage check: list each feature and ask, “Would I know this at the time of inference?” If not, it’s a trap—and AI-900 loves this trap because it tests understanding, not memorization.
1. A retail company has historical data that includes customer demographics and whether each customer responded to a previous email campaign (Yes/No). The company wants to predict whether a new customer will respond to the next campaign. Which machine learning approach should you use?
2. You train a model in Azure Machine Learning to predict house prices. It performs very well on the training dataset but significantly worse on new, unseen data. Which term best describes this situation?
3. A bank builds a binary classification model to detect fraudulent transactions. Fraud is rare, and the bank wants to minimize missed fraud cases. Which evaluation metric is most appropriate to prioritize?
4. You are preparing data for a supervised learning experiment in Azure Machine Learning. Which split is primarily used to make a final, unbiased evaluation of how well the trained model will perform on new data?
5. A logistics company wants to automatically group delivery addresses into segments based on similarity (for example, to design regional routes) but does not have predefined labels for the segments. Which machine learning workload best fits this requirement?
This chapter maps directly to the AI-900 objective: Identify computer vision workloads on Azure. The exam expects you to recognize common vision scenarios (image analysis, image classification, object detection, OCR) and select the most appropriate Azure capability for a given business need. Because AI-900 is fundamentals-focused, you are not tested on writing code or tuning deep learning models—but you are tested on choosing the correct service and explaining what it does in plain language.
You’ll see scenario prompts like “analyze photos,” “extract text,” or “process invoices.” The trick is to translate those words into the correct workload type and then into the right Azure service family. You also need to show awareness of responsible AI in vision: privacy, consent, and bias can appear as qualifiers in a question stem and change the best answer.
In this chapter, you’ll practice identifying vision scenarios (classification, detection, OCR), selecting the right Azure vision capability, and understanding document/form processing at a high level. The final section focuses on how to pick the best service from a scenario—the core exam skill.
Practice note for Identify vision scenarios: image classification, object detection, and OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Select the right Azure vision capability for a given use case: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand document and form processing at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Computer vision (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify vision scenarios: image classification, object detection, and OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Select the right Azure vision capability for a given use case: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand document and form processing at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: Computer vision (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Identify vision scenarios: image classification, object detection, and OCR: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Select the right Azure vision capability for a given use case: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-900’s computer vision objective is about recognizing what kind of vision problem you have and which Azure service category can solve it. At a high level, the exam separates (1) understanding images (what’s in the picture), (2) locating things in images (where something is), and (3) reading text from images and documents (OCR and extraction). You should be able to map each to a business-friendly scenario.
Azure groups these capabilities primarily under Azure AI services. In exam language, you will commonly see references to “prebuilt models” versus “custom models.” Prebuilt models are ready-to-use for common tasks (for example, describing an image, detecting common objects, or extracting text). Custom models are used when you must recognize domain-specific categories (for example, types of defects on a circuit board) that a generic model is unlikely to cover.
Exam Tip: When a question emphasizes “no training data,” “quickly add vision,” or “out-of-the-box,” it’s steering you toward a prebuilt capability. When it emphasizes “your own categories,” “brand-specific products,” or “specialized objects,” it’s steering you toward custom vision/model training.
Common trap: confusing “image analysis” with “object detection.” Image analysis often returns tags, captions, categories, and general insights. Object detection implies bounding boxes (location) and is often used for counting or tracking items in a scene. If the scenario needs coordinates or “find all instances of X,” think detection.
Start every vision question by identifying the workload type. Three appear repeatedly in AI-900: image classification, object detection, and OCR. The exam may also describe “image analysis” tasks like tagging and captioning, which are often prebuilt.
Exam Tip: If the prompt includes verbs like “count,” “locate,” “track,” “find all,” or “draw boxes,” it’s object detection. If it includes “classify as,” “categorize,” or “label,” it’s classification. If it includes “extract text,” “read,” “digitize,” or “scan,” it’s OCR/document processing.
Common trap: assuming OCR is enough for “invoice processing.” OCR extracts text, but invoice processing often requires structure (vendor name, invoice number, line items). That usually points to a document extraction capability designed for forms, not just raw text recognition.
AI-900 frequently tests your ability to choose between prebuilt image analysis and custom model approaches. The simplest decision point is: do you need generic understanding of common things, or do you need to recognize your own domain-specific categories?
Image analysis (prebuilt): best when you want broad insights such as describing an image, generating tags, detecting common objects, or identifying scenes. It’s ideal for rapid prototyping and for scenarios where “good enough” general recognition is acceptable.
Custom vision (custom model): best when you need to classify or detect items that are specific to your organization or niche—such as a particular product packaging variant, a proprietary component, or a set of defect types that don’t exist in standard datasets. This approach requires labeled images and a training/evaluation cycle, even if the platform abstracts much of the ML complexity.
Exam Tip: Watch for constraints in the stem: “limited time,” “no ML expertise,” and “no labeled dataset” lean toward prebuilt analysis. “High accuracy on specific items,” “unique classes,” and “you can provide labeled images” lean toward custom.
Common trap: picking custom vision just because the scenario mentions “images.” The exam often rewards the simplest workable solution. If the task is “generate tags for photos” or “detect whether an image contains a dog,” prebuilt analysis is usually the intended answer—no need to train a custom model.
Another trap is conflating “classification” and “detection” when talking about custom models. Custom vision can support both, but if the scenario asks for the location of each instance, the correct workload is object detection, not classification.
OCR questions on AI-900 range from simple “read text in an image” to more business-driven “process documents.” The exam expects you to understand the difference between raw text recognition and structured document extraction.
OCR basics: OCR converts text in images into machine-readable characters. Use cases include reading signs in photos, extracting text from screenshots, or digitizing paper documents. If the question only asks to “extract all text,” OCR alone is sufficient.
Document and form processing: Many business scenarios require you to extract fields (like invoice number, date, total) and sometimes tables (line items). This is more than OCR because you need to understand layout and map text to named fields. Azure provides capabilities for analyzing documents such as invoices, receipts, business cards, and IDs using prebuilt models, and can also support training for custom document types.
Exam Tip: When the prompt mentions “receipts,” “invoices,” “forms,” “contract fields,” “key-value pairs,” or “line items,” it’s usually testing whether you know that document extraction is a specialized workload—not just OCR. The best answer often references a document intelligence/form recognition capability rather than generic text extraction.
Common trap: choosing an image analysis service for document processing because it “handles images.” Document extraction is assessed on accuracy of field capture and structure, not on image tagging/captioning. Another trap is assuming you must build a custom model for every form; if the document type is common (receipt/invoice/ID), prebuilt models are frequently the intended choice.
AI-900 includes Responsible AI concepts across all workloads, and computer vision is a frequent area for privacy and fairness concerns. The exam may embed ethical qualifiers in a scenario (for example, “public space,” “employees,” “children,” “medical images”) and expect you to identify risk and a safer approach.
Privacy and consent: Images can contain personally identifiable information (PII) such as faces, license plates, addresses, and ID numbers. A responsible solution minimizes data collection, uses access controls, retains data only as long as necessary, and obtains appropriate consent and legal basis for processing.
Bias and fairness: Vision models can perform differently across demographics or conditions (lighting, skin tones, cultural attire). If a scenario involves decisions that impact people (security screening, hiring, access control), the exam expects awareness that model evaluation must include diverse data and ongoing monitoring.
Transparency and accountability: Stakeholders should know when automated vision is being used, what it’s for, and how errors are handled. Human review is often appropriate for high-stakes outcomes.
Exam Tip: If a question asks “what should you do first” or “best practice,” look for answers about consent, data minimization, evaluating bias, and human oversight—not technical performance tweaks. Fundamentals exams often reward governance steps over implementation details.
Common trap: treating responsible AI as optional. On AI-900, responsible AI is part of the expected reasoning. If two services could work, the scenario’s privacy constraints may determine the best choice (for example, preferring to extract only needed fields rather than storing full images).
On the AI-900 exam, “select the right service” is often the real task. Use a repeatable approach: (1) identify the workload type, (2) decide prebuilt vs custom, (3) match to the Azure capability, and (4) validate against constraints like speed, data availability, and governance.
Exam Tip: When multiple answers sound plausible, choose the one that most directly matches the deliverable in the scenario (tags, boxes, extracted fields) with the least additional work. AI-900 scenarios typically favor managed, prebuilt services unless the stem explicitly demands customization.
Common trap: overfitting to brand names in the options. Some questions list several Azure services; focus on the capability described, not the most familiar name. Also watch for “document” wording: if the source is a scanned invoice or receipt, the exam is usually checking whether you can distinguish document extraction from generic OCR.
1. A retail company wants to automatically tag product photos with labels such as "shoe", "hat", and "backpack" to improve search results on their website. The company does not need the location of items in the image—only the best label for each image. Which computer vision scenario is this?
2. A city wants to analyze traffic-camera images to identify and draw bounding boxes around vehicles and pedestrians in each frame. Which type of computer vision workload best fits this requirement?
3. An insurance company receives scanned claim forms and wants to extract structured fields such as policy number, claim amount, and claimant name from each document. Which Azure AI capability is the best fit?
4. A media company wants to extract text from images of event posters and store the text for search and indexing. Which computer vision scenario is being described?
5. A company plans to analyze photos uploaded by users to detect whether the images contain faces. Because the photos are user-generated, the company wants to reduce privacy risk and meet responsible AI expectations. Which action is most appropriate?
This chapter targets the AI-900 objectives that ask you to identify Natural Language Processing (NLP) workloads and generative AI workloads on Azure, and to choose the right service and approach for common business scenarios. As a non-technical pro, you are not expected to code models, but you are expected to recognize patterns: when classic NLP is sufficient (sentiment, entities, translation, summarization) versus when generative AI is the better fit (drafting, Q&A over documents, rewriting, multi-step reasoning).
The exam often tests selection logic more than definitions: “Given this scenario, which Azure capability should you use?” You’ll practice that logic throughout this chapter by mapping scenarios to (1) text analytics features, (2) conversational and speech experiences, and (3) Azure OpenAI / generative AI approaches. You’ll also see how Responsible AI shows up as practical guardrails (safety, grounding, content moderation), not just theory.
Exam Tip: When a question emphasizes “extract,” “detect,” “classify,” or “translate,” think classic NLP. When it emphasizes “generate,” “draft,” “summarize in a specific style,” “answer with reasoning,” or “write code,” think generative AI (often via Azure OpenAI). Many wrong answers on AI-900 are “too powerful” (GenAI) for a simple extraction need, or “not powerful enough” (classic NLP) for an open-ended generation task.
Practice note for Understand NLP scenarios: sentiment, entities, summarization, translation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize conversational AI patterns and speech scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI fundamentals and prompt basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: NLP + Generative AI (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scenario workshop: choose between classic NLP and GenAI: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand NLP scenarios: sentiment, entities, summarization, translation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Recognize conversational AI patterns and speech scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Explain generative AI fundamentals and prompt basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Practice set: NLP + Generative AI (exam-style questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scenario workshop: choose between classic NLP and GenAI: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
NLP workloads focus on understanding and working with human language in text (and sometimes speech transcribed into text). For AI-900, you should recognize the most common NLP scenarios: sentiment analysis on customer feedback, entity recognition in documents (names, organizations, locations), key phrase extraction to find themes, language detection and translation, and summarization for long content.
On Azure, these capabilities are commonly delivered through Azure AI services for language features (often referred to in study materials as “Language” capabilities). The exam expects you to connect a scenario to a capability, not to memorize every API name. Read the stem carefully: “classify” usually means assign labels (e.g., positive/negative or categories), “extract” means pull out structured bits from unstructured text, and “summarize” means reduce length while preserving meaning.
Also recognize the boundary between rules and AI. If the scenario is deterministic and easy to encode (e.g., “if the email contains ‘unsubscribe’ then label it opt-out”), a rule can be enough. If you need to interpret nuanced language (“I expected better, but support was kind”), AI is more appropriate.
Exam Tip: If the question mentions “structured output from unstructured text” (entities, key phrases), classic NLP is typically the most direct and cost-effective choice. Don’t over-select generative AI unless the scenario requires free-form output or creative rewriting.
Common trap: confusing document processing (OCR) with NLP. OCR converts images to text (a vision workload); NLP then analyzes that text. Many scenarios in business use both, but the exam will often focus on one step at a time.
Text analytics is a bucket of classic NLP tasks that turn text into insights. The exam frequently frames these as “analytics” rather than “generation.” Three patterns appear repeatedly: classification, extraction, and summarization.
Classification assigns labels. Examples include sentiment (positive/negative/neutral), topic categories (billing issue vs product defect), or intent-like routing (sales lead vs support request). The key is that the output is one or more categories, not a rewritten paragraph.
Extraction pulls structured data from text. Entity recognition finds items like people, organizations, places, dates, and product names. Key phrase extraction identifies the main talking points. This is ideal when downstream systems need fields, tags, or highlights rather than prose.
Summarization condenses content. On AI-900, summarization can appear as an NLP scenario (produce a shorter version of the same content). Pay attention to constraints: if the question demands a specific tone (“executive summary in 3 bullet points”) or asks you to incorporate outside context, that leans toward generative AI. If it’s “summarize this article,” classic summarization fits.
Exam Tip: When a question asks for “key information to index/search,” think extraction (entities/key phrases). When it asks for “overall mood,” think sentiment. When it asks for “convert Spanish to English,” think translation—don’t choose generative AI just because it can also translate.
Common trap: treating “summarization” as always generative. The exam may present summarization as a classic NLP capability. Your clue is whether the output must be tightly controlled (length only) or creatively adapted (style, persona, compliance formatting). Adaptation and style usually indicate GenAI.
Conversational AI combines language understanding with an interaction pattern: a user says or types something, the system responds, and context may carry across turns. For AI-900, you should recognize two core building blocks: speech scenarios (speech-to-text and text-to-speech) and chat experiences (bots and conversational interfaces).
Speech-to-text converts spoken audio into written text. This is used for call center transcription, meeting notes, and voice commands. Text-to-speech generates spoken audio from text, useful for accessibility and voice assistants. In many real solutions, speech-to-text feeds an NLP step (sentiment or entity extraction) to analyze calls, or feeds a GenAI step to summarize a meeting.
Chat experiences come in two major patterns you should be able to identify: (1) a scripted or guided bot that routes users through options, and (2) a more flexible assistant that handles open-ended questions. The exam often tests which approach is appropriate. If the business wants consistent, controlled answers (“store hours, return policy”), a structured conversational experience is a safer fit. If it wants broad help (“draft an email to a customer about…”) then a generative assistant is more suitable.
Exam Tip: If the prompt mentions audio input/output, your first mental model should be speech services (transcription or voice synthesis). If it mentions multi-turn interaction, remember that “conversation” is a pattern and can be implemented with classic NLP, GenAI, or both.
Common trap: assuming a chatbot must be generative. Many chatbots are retrieval/FAQ or decision-tree based. Look for wording like “predictable responses,” “approved answers,” or “compliance.” Those phrases signal you should avoid open-ended generation unless guardrails are explicitly included.
Generative AI workloads produce new content: text, summaries, drafts, explanations, or even code. On Azure, this commonly maps to Azure OpenAI, where large language models (LLMs) can be used to build assistants, copilots, and content generation features. For AI-900, you must distinguish GenAI from classic ML and classic NLP: GenAI is optimized for generating language, while classic NLP is optimized for extracting or classifying language.
Typical GenAI use cases tested on the exam include: drafting emails and marketing copy, summarizing long documents into a tailored format, generating FAQs from internal documentation, and building Q&A assistants over enterprise content. The exam often frames this as “foundation models” that can be adapted with prompting or augmented with your data.
Key concept: GenAI responses are probabilistic and can be incorrect. That’s why questions about enterprise assistants often include requirements like “use my company documents” or “cite sources.” Those phrases are signals to use grounding approaches (retrieval-augmented generation) rather than letting the model answer purely from its general training.
Exam Tip: If a scenario asks for “human-like response,” “drafting,” “rewriting,” “brainstorming,” or “natural language Q&A,” default to generative AI. If it asks for “extract fields,” “detect sentiment,” or “classify,” default to classic NLP. The correct answer is usually the simplest tool that meets the requirement.
Common trap: picking GenAI when the question wants deterministic results. For example, compliance tagging or extracting invoice numbers is better served by extraction and rules, not a model that might paraphrase or hallucinate numbers.
Foundation models are large pre-trained models that can be applied to many tasks without training from scratch. In exam language, this is why GenAI can perform summarization, translation, classification-like tasks, and question answering through prompting. A copilot is an application pattern: an assistant embedded into a workflow (CRM, HR portal, help desk) that helps users produce or find information.
Prompting is how you steer the model. Two prompt styles show up frequently in AI-900 prep: zero-shot and few-shot. Zero-shot means you give instructions with no examples (“Summarize this complaint in one sentence and tag sentiment.”). Few-shot means you include examples of input/output to teach the format and expectations (“Example 1… Example 2… Now do this one.”). Few-shot prompting often improves consistency, especially for structured outputs.
When reading scenario questions (including the chapter practice set you’ll encounter elsewhere in the course), identify whether the goal is format consistency or creativity. If the system must always output JSON-like fields, bullet lists, or specific labels, that’s a clue to use clear constraints, examples (few-shot), and explicit formatting instructions.
Exam Tip: “Few-shot” is a favorite term because it sounds like training but it isn’t. It’s still prompting. If an option says “train a custom model” when the scenario only needs better formatting, that’s often the trap.
Common trap: confusing few-shot prompting with fine-tuning. Fine-tuning changes model behavior through additional training; few-shot adds examples in the prompt at runtime. AI-900 typically emphasizes the conceptual difference rather than deep implementation steps.
Responsible AI is tested across AI-900, but generative AI increases the importance because the system can produce unsafe, biased, or incorrect content. For exam purposes, focus on three practical concepts: safety controls, grounding, and content moderation.
Safety refers to preventing harmful outputs (hate, violence, self-harm instructions, sexual content, or disallowed advice). Safety also includes reducing prompt injection risks (users trying to override instructions) and limiting data exposure. Expect scenario phrasing like “must not generate offensive content” or “must comply with company policy.” Those are direct signals that safety and moderation features are required.
Grounding means anchoring responses in trusted data (your organization’s documents, databases, or approved knowledge base) instead of letting the model answer purely from its general knowledge. Grounding reduces hallucinations and helps with enterprise requirements like “answer only using our policy documents.” Some questions will hint at this with language like “must cite sources” or “must be based on internal manuals.”
Content moderation is the mechanism that detects and filters unsafe user inputs and model outputs. On Azure, moderation capabilities are commonly discussed as part of building safe GenAI apps. The exam cares that you know moderation exists and why you would apply it, not that you memorize threshold settings.
Exam Tip: In scenario workshops (choose between classic NLP and GenAI), a frequent deciding factor is risk. If the organization requires controlled, auditable outputs, prefer classic NLP or grounded GenAI with moderation over “open-ended” generation.
Common trap: assuming grounding eliminates the need for moderation. Even grounded systems can produce problematic wording or be attacked through prompts. Another trap is assuming GenAI is always the right choice for summarization—if the requirement is compliance-critical, classic summarization plus extraction may be safer and more consistent.
1. A retail company wants to automatically identify whether customer reviews are positive, negative, or neutral to track product satisfaction trends. Which Azure AI capability is the best fit?
2. A legal team needs to extract names of people, organizations, and locations from thousands of contract documents so they can populate a searchable index. Which approach should you choose?
3. A support center wants an interactive voice bot that allows customers to speak their issue and receive spoken responses. Which set of Azure capabilities best matches the scenario?
4. A marketing team wants to generate a first draft of a product announcement in a specific brand voice and then rewrite it for different audiences (executives vs. end customers). Which Azure service is most appropriate?
5. A company has an internal knowledge base (policies, HR docs, and FAQs). Employees want a chat experience that answers questions using only the company documents and avoids making up information. What is the best approach?
This chapter is the capstone for AI-900: you’ll simulate the exam experience, pressure-test your understanding across all objectives, and convert “I’ve read it” knowledge into “I can answer it” performance. AI-900 rewards clear scenario matching: identify the workload (AI vs rules, ML vs vision vs NLP vs GenAI), pick the right Azure service family, and confirm responsible AI considerations. The goal of the mock exam parts is not to “score once,” but to create an evidence-based map of what you miss, why you miss it, and what wording patterns the exam uses.
You’ll complete two mixed-domain mock sets (Part 1 and Part 2), then do a Weak Spot Analysis using a repeatable answer-review method. Finally, you’ll run an exam-day checklist so the only surprises you face are the ones inside the questions.
Exam Tip: Many AI-900 wrong answers are “almost-right services.” Learn to spot what the scenario is actually asking (classification vs object detection vs OCR, extractive summarization vs generative text, rules-based automation vs predictive ML), then choose the simplest service that meets the requirement.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Treat each mock exam like a rehearsal for the real thing: one sitting, no notes, no searching, and no pausing for deep reading. Your goal is to build the skill the exam tests most: fast, accurate workload identification under time pressure. Set a timer and commit to a pacing plan that leaves review time at the end. Even though AI-900 is fundamentals, candidates often run out of time because they overthink “familiar” scenarios.
Use a three-pass approach. Pass 1: answer everything you can confidently and flag anything with uncertainty. Pass 2: return to flagged items and eliminate distractors using scenario keywords (data type, output, constraints, and whether training is required). Pass 3: final sweep for misreads—especially “best,” “most cost-effective,” “no-code,” “real-time,” “without training,” or “must explain.”
Exam Tip: On AI-900, “without training” often points to prebuilt AI services (e.g., Vision, Language, Speech) rather than Azure Machine Learning. “Must train a model” generally points to ML workflows and evaluation.
Finally, store your results by objective domain (AI workloads, ML, vision, NLP, GenAI). The exam blueprint is broad; your review must be targeted. A raw score alone doesn’t tell you what to fix—patterns do.
Mock Exam Part 1 (Set A) should feel like the early-to-mid portion of the real exam: straightforward scenarios with a few traps. Focus on quickly classifying the workload. Start by asking: What is the input (text, images, audio, tabular data)? What is the output (label, score, extracted text, generated content)? And do we need training, or can we use a prebuilt capability?
Common scenario families in Set A include: (1) “Choose AI vs rules.” If the logic is static and fully described, rules are fine; if it requires pattern recognition from data (fraud, churn, sentiment), AI/ML fits. (2) “Select the right service.” This is usually about the Azure AI service families: Vision for image analysis/OCR, Language for text analytics and conversational language, Speech for transcription and voice, and Azure OpenAI for generative tasks. (3) “Responsible AI.” Look for fairness, transparency, privacy, and reliability cues—especially when dealing with sensitive attributes or decisions that affect people.
Exam Tip: If the scenario says “extract key phrases,” “detect sentiment,” or “recognize entities,” that’s classic NLP text analytics (Azure AI Language). If it says “write,” “summarize creatively,” “draft,” or “generate,” it’s generative AI (Azure OpenAI). Extractive vs generative is a frequent separator on AI-900.
After completing Set A, don’t just check correctness—note whether misses came from terminology (service names), concept confusion (training vs inference, classification vs regression), or reading error (skipped a constraint like “near real-time” or “no code”).
Mock Exam Part 2 (Set B) should increase difficulty by combining requirements and introducing “best choice” trade-offs. Expect scenarios that mix multiple AI workloads, such as a customer support system that needs sentiment detection (NLP), speech-to-text (Speech), and a chatbot (conversational AI), or a document workflow that needs OCR (Vision) plus entity extraction (Language). The exam often tests whether you can separate a multi-step solution into the correct service components.
Set B also tends to probe fundamental ML concepts: training vs evaluation, overfitting vs generalization, and what metrics imply. While you won’t be doing math, you will be asked to reason: Is it classification or regression? Do we need labeled data? What indicates a model may be overfit (excellent training performance, poor test performance)?
Exam Tip: If a scenario requires predicting a numeric value (price, demand), it’s regression. If it’s choosing a category (approve/deny, churn/no churn), it’s classification. If it’s grouping without labels, it’s clustering. The exam frequently hides this behind business language.
Finish Set B with the same discipline: log misses by objective domain and label the reason. This will drive your Weak Spot Analysis rather than guessing what to study next.
Your review process determines your score improvement more than the number of mock questions you take. Use a consistent framework to explain (1) why the correct answer is correct, and (2) why each distractor is wrong. The exam’s distractors are designed to test surface familiarity—so your job is to attach each choice to a specific capability boundary.
Apply the “Four-Box Review” to every missed or flagged item:
Exam Tip: When two answers seem plausible, ask: “Which option is the minimum solution that satisfies all constraints?” AI-900 typically favors the most direct service match, not the most configurable platform.
Common distractor patterns to call out in your notes: (1) “Wrong domain” (Vision vs Language). (2) “Wrong task within domain” (classification vs detection vs OCR). (3) “Overkill platform” (choosing Azure Machine Learning when a prebuilt service is explicitly enough). (4) “Wrong AI type” (generative vs extractive). (5) “Violates constraint” (requires training, requires labeled data, or doesn’t support real-time needs).
Finish review by converting each mistake into a one-sentence rule you can reuse. Example: “If the task is to read printed text in images, pick OCR; if it’s to identify what the image contains, pick image analysis/classification.” These rules become your last-day revision sheet.
Use this final checklist to ensure coverage of the course outcomes and the AI-900 objective style. You’re not memorizing product marketing—you’re building quick matching rules between scenario language and solution category.
Exam Tip: If you’re shaky on a domain, don’t re-read everything. Rework your own “if-then” rules and test them against mixed scenarios. AI-900 performance comes from fast categorization.
This checklist aligns directly to the exam’s coverage: describe workloads, select the right Azure AI capability family, and reason about responsible use. Your Weak Spot Analysis should now point to one or two domains to polish, not five to re-learn.
Exam-day performance is a systems problem: environment, timing, and composure. Start with your setup. If remote, ensure a quiet room, clear desk, stable internet, and allowed identification ready. If in a test center, arrive early and expect check-in time. These are not “extras”—they protect your mental bandwidth for scenario interpretation.
Use a simple confidence routine: before starting, remind yourself of your decision tree—identify modality (text/image/audio/tabular), choose workload type, apply constraints, then eliminate distractors. This keeps you from drifting into overthinking.
Exam Tip: If you feel stuck, stop rereading and instead restate the requirement in your own words: “The business wants X output from Y input under Z constraint.” Then pick the service family that does exactly that. This breaks the loop of indecision.
Finally, run a last-minute “two-minute review” of your personal rules: classification vs regression, OCR vs detection, extractive vs generative, prebuilt services vs custom ML, and responsible AI principles. When you sit down, your goal is steady execution—not new learning. If you’ve completed both mock exam parts, performed a structured review, and addressed weak spots, you’re prepared to score consistently.
1. A company wants to extract printed text from photos of delivery receipts and store the extracted text in a database. Which Azure AI capability should you use?
2. You are designing a solution that must generate a customer-support reply in natural language based on a user’s question and the conversation history. Which approach best fits this requirement?
3. A team needs to predict whether a customer will cancel a subscription (Yes/No) using historical customer data (usage, tenure, support tickets). Which workload type is this, and what is the most appropriate Azure service to build it?
4. You build a face-based access solution and want to follow Microsoft’s responsible AI principles. Which action best addresses the principle of fairness?
5. During a weak-spot review, you notice you frequently confuse extractive summarization with generative summarization. On the AI-900 exam, which statement best describes extractive summarization?