AI Certification Exam Prep — Beginner
Master AI-102 with objective-mapped practice tests, rationales, and a mock exam.
This course is a practice-test-first blueprint for learners targeting the Microsoft AI-102 (Azure AI Engineer Associate) certification. If you’re new to certification exams but have basic IT literacy, you’ll get a clear path from exam registration through final review—while building confidence with more than 250 exam-style questions mapped directly to Microsoft’s official skills domains.
AI-102 measures your ability to design, implement, and operate AI solutions on Azure. That includes selecting the right Azure AI services, implementing generative AI patterns responsibly, building agentic experiences that can use tools safely, and delivering real-world computer vision, natural language, and search-based information extraction solutions.
Chapter 1 gets you set up: how to register, what to expect on exam day, how AI-102 is scored, and how to study efficiently using objective mapping. Chapters 2–5 each focus on one or two domains and combine targeted explanations with exam-style question sets designed to surface real exam pitfalls (multi-select logic, scenario constraints, best-answer wording, and service selection tradeoffs). Chapter 6 finishes with a full mock exam experience, weak-spot analysis, and a final review plan you can run in the last week before your scheduled date.
Every practice block is designed to reinforce decision-making: which Azure AI service fits, how to secure it, how to troubleshoot it, and how to justify it. You’ll also train the “exam muscle” of reading carefully, eliminating distractors, and managing time across mixed question types.
Use this course as your primary practice engine alongside Microsoft Learn or your preferred notes. If you’re ready to begin, Register free and start tracking your progress by domain. You can also browse all courses to pair AI-102 with complementary Azure fundamentals content.
This course is built for beginners to certification testing who want a structured, domain-aligned way to prepare for AI-102. If you can navigate cloud concepts, understand APIs at a high level, and are willing to practice consistently, you’ll have a clear plan from your first diagnostic set to your final mock exam.
Microsoft Certified Trainer (MCT) — Azure & AI
Jordan Patel is a Microsoft Certified Trainer (MCT) who designs Azure and AI certification prep for early-career and transitioning IT professionals. He specializes in mapping hands-on Azure AI services to Microsoft exam objectives and turning common exam traps into repeatable study strategies.
AI-102 is not a “vocabulary” exam; it is a solution-design-and-implementation exam focused on Azure AI services and how you operate them in the real world. Expect questions that blend architecture choices (which service or feature fits), implementation details (SDK vs REST, parameters, indexing pipelines), and operational responsibility (security, monitoring, cost control, deployment). This chapter sets your compass for the whole course: how the exam is structured, what skills it measures, and how to turn practice tests into a deliberate improvement loop rather than random repetition.
Your goal is to walk into the exam with three advantages: (1) you can map any question to an objective domain quickly, (2) you recognize common traps that target partial knowledge, and (3) you have a repeatable study system that converts misses into durable skill. The sections below align directly to the lessons in this chapter: exam format and question styles, registration and testing logistics, scoring and pitfalls, a beginner-friendly plan using objective mapping, and an effective review-and-retake strategy.
Practice note for Understand the AI-102 exam format and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register, schedule, and prep your testing environment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scoring, passing criteria, and how to avoid common pitfalls: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Set up a beginner-friendly study plan using objective mapping: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for How to review practice tests: rationales, notes, and retake strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Understand the AI-102 exam format and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Register, schedule, and prep your testing environment: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Scoring, passing criteria, and how to avoid common pitfalls: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Set up a beginner-friendly study plan using objective mapping: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for How to review practice tests: rationales, notes, and retake strategy: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-102 measures whether you can plan, implement, and operate an Azure AI solution end-to-end. Use the course outcomes as your north star: resource planning and governance; generative AI with Azure OpenAI (prompting, RAG, safety, evaluation); agentic solutions (tool use, orchestration, grounding, state); computer vision (image analysis, OCR, document intelligence); NLP (text analytics, conversation, translation); and knowledge mining via Azure AI Search (indexing, enrichment, extraction).
On the exam, “domain mapping” is how you stop guessing. When you read a question, immediately label it: is this primarily governance/ops (RBAC, private endpoints, monitoring, cost), genAI (prompting, embeddings, RAG, content filters), agents (tools/functions, orchestration patterns, memory/state), vision (OCR vs document models), NLP (language detection, sentiment, PII, conversational patterns), or search/knowledge mining (indexers, skillsets, chunking, vector search)? That label tells you which details matter and which are distractors.
Exam Tip: The exam often includes realistic constraints—data residency, network isolation, latency, cost ceiling, or “must not store prompts.” These are not flavor text; they are the requirements that determine the correct service configuration. Underline constraints and map each one to a control (e.g., private link for network, managed identity for auth, content filtering/evaluations for safety).
Common trap: choosing a “popular” service instead of the one that matches the workload. For example, OCR needs are not all the same: basic image OCR vs structured document extraction implies different tooling. Likewise, RAG is not “use search”; it is retrieve the right chunks with the right ranking (keyword + vector, filters, freshness) and ground responses with citations.
Registering correctly prevents last-minute issues that cost points before you even start. You’ll schedule through Microsoft’s certification portal and take the exam through Pearson VUE (either online proctored or at a test center). Build a checklist and complete it early: verify your legal name matches your ID, confirm the correct exam code (AI-102), and choose a time slot when you are mentally sharp (not “after work if possible”).
Online proctoring logistics are strict. Your testing space must be clean, quiet, and compliant; your webcam must show your face and workspace. Close background apps and disable notifications. Run Pearson VUE’s system test days ahead, not minutes before. If you have multiple monitors, know whether you must disconnect them. The exam is challenging enough—don’t add avoidable friction.
Exam Tip: Treat your testing environment like a deployment: do a “pre-flight” validation. Test internet stability, power, audio, and camera positioning. Keep your ID ready and ensure lighting is adequate. A technical interruption can break focus and time management.
Accommodations: if you need extra time or assistive arrangements, request them well in advance. Don’t assume you can “sort it out” on exam day. Also decide whether a test center is better for you (more stable environment, fewer home interruptions) versus online (convenience). Pick the option that maximizes reliability and minimizes cognitive load.
Microsoft exams are scored on a scaled model; you don’t need a perfect percentage to pass, but you do need consistent performance across objectives. Questions can be single-answer, multi-select, ordering, drag-and-drop matching, and longer scenario-based sets (often called case studies). The key performance skill is not memorization—it’s applying requirements to the correct configuration or design choice.
Expect case studies where multiple questions share the same scenario and constraints. Treat the scenario as a mini design brief: capture requirements, constraints, and existing architecture before you answer anything. Many candidates lose points by answering the first question too quickly and then realizing later that a constraint (e.g., “no public endpoint,” “customer-managed keys,” “must log prompts”) should have influenced all answers.
Labs/simulations vary by exam delivery and current format. You should be prepared for “implementation-flavored” questions that feel like a lab even when they’re multiple choice: choosing CLI/portal steps, selecting correct API parameters, or identifying which setting enforces network isolation. This course’s practice tests are designed to build that muscle: not just what a feature is, but what you would do to enable it.
Exam Tip: When a question describes a failure (“403,” “throttling,” “index not updating,” “model output is unsafe”), you are often being tested on the operational control plane: keys vs managed identity, quotas, monitoring, indexer schedules, content filters, evaluation gates. Don’t jump straight to “change the model” unless the symptom truly points there.
Time management is a skill you can train. Multi-select questions are the biggest time sink because they invite second-guessing. Your job is to convert them into a deterministic process: (1) restate the requirement in your own words, (2) eliminate options that violate constraints, (3) select only what is necessary and sufficient. If the prompt says “choose two,” treat it like a contract—over-selecting is not “safer,” it’s wrong.
Scenario questions require disciplined reading. Skim first for what’s being asked, then read the scenario to extract only relevant facts: security boundaries, data sources, latency needs, cost constraints, update frequency, and compliance requirements. Write a mental “spec” and keep it consistent. A common trap is being distracted by technical detail that is irrelevant to the decision (e.g., programming language mentioned purely as context).
Exam Tip: Use a two-pass strategy. Pass 1: answer straightforward items quickly, mark uncertain ones. Pass 2: return to marked questions with the remaining time. This reduces the risk of spending 6 minutes on a single multi-select early and then rushing through an entire case study later.
Watch for “absolutes” in options: words like always, never, or “the only way” are frequently (not always) incorrect on Azure exams because Azure usually provides multiple valid approaches; the question is which is best given constraints. Your goal is to identify the best fit, not merely something that could work.
Practice tests only work if you review them like an engineer, not like a gambler. The study loop for AI-102 should be: Learn → Quiz → Remediate → Retest. “Learn” means targeted reading and minimal hands-on exposure to the feature set (enough to understand the knobs). “Quiz” means objective-mapped practice questions under time constraints. “Remediate” means you write down why each wrong answer was wrong and what signal you missed. “Retest” means you reattempt after a delay so you’re testing memory and reasoning, not recall of the option letters.
When reviewing rationales, do not stop at “what is correct.” Identify the decision rule. Example decision rules: “Use Azure AI Search indexers for scheduled ingestion from supported sources,” “Use managed identity over API keys when the requirement says ‘no secrets’,” “Use content filters/safety system messages when the requirement is to prevent unsafe output,” “Use vector + hybrid retrieval when semantic similarity is required.” Those rules become your mental shortcuts on exam day.
Exam Tip: Create an “error taxonomy.” Tag every miss as one of: misunderstood requirement, confused services, missed security constraint, misread multi-select, or operational gap (monitoring/cost/deployment). Over time, your tags will reveal the real reason you’re losing points—most candidates discover it’s not lack of knowledge, but pattern errors like skipping constraints or over-selecting.
Retake strategy: don’t immediately redo the same set. Wait 24–72 hours, then retest. Mix topics to simulate the exam’s interleaving. Your goal is stable performance, not a high score that comes from memorizing yesterday’s answer key.
An objective tracker turns a large syllabus into a measurable plan. Build a simple spreadsheet or note system with rows for each official exam domain and sub-skill (planning/management, generative AI, agentic patterns, vision, NLP, search/knowledge mining). For each row, track: confidence (1–5), last practiced date, common errors, and “next action” (read docs, do a mini lab, take a question set).
Map every practice test question you miss back to exactly one objective row. If you can’t decide where it belongs, that’s a signal you don’t yet understand the boundaries between services—an exam-relevant weakness. Over time, you should see clusters: perhaps you’re strong in vision and NLP but weak in Azure AI Search enrichment pipelines; or you know prompting but struggle with safety/evaluation and deployment governance.
Exam Tip: Include operational controls as first-class items in the tracker. Many candidates track only “features,” but AI-102 frequently tests production readiness: RBAC, managed identities, network isolation, monitoring, quotas, and cost controls. Add columns for “security control needed?” and “monitoring signal?” so you practice thinking like the owner of the system.
Finally, connect the tracker to your calendar. Assign weekly blocks by objective weight and your weak areas. Keep the loop tight: learn the minimum, test quickly, remediate deeply, and retest until your tracker shows consistent confidence across domains. That is how you turn 250+ objective-mapped questions into a passing score and real job-ready capability.
1. You are creating a study plan for AI-102. You want a repeatable way to decide which Azure AI service area a practice question is testing before you look at the answer. Which approach best aligns to the AI-102 exam’s objective-based design?
2. A team consistently scores well on knowledge checks but misses exam questions that include subtle constraints (for example, cost control, security boundaries, deployment operations). Which pitfall is most likely causing the issue, and what is the best correction strategy?
3. You are reviewing a practice test question you got wrong. You want the highest-impact review process to reduce repeat mistakes on similar items (for example, choosing SDK vs REST, selecting the right indexing pipeline, or applying monitoring/cost controls). What should you do next?
4. A company wants to ensure candidates are prepared for the AI-102 exam’s question styles. Which statement best reflects what they should expect during the exam based on the chapter summary?
5. You are coaching a beginner who feels overwhelmed by the breadth of AI-102 topics. You want a study approach that scales and prevents gaps from being hidden by overall practice test scores. Which plan is most aligned with the chapter’s recommended ‘game plan’?
AI-102 does not only test whether you can call an API; it tests whether you can run an AI solution in Azure like an engineer: choosing the right service, deploying it to the right region, securing it correctly, and operating it reliably under cost and compliance constraints. This chapter maps directly to the “plan and manage” expectations that show up in scenario questions—especially those that force trade-offs between security, latency, availability, and spend.
Expect prompts that include constraints such as “data must not traverse the public internet,” “only specific app components may access the model,” “must support multiple environments,” “must meet budget,” or “must be deployable in a region where the service is available.” Your job on the exam is to translate those constraints into concrete Azure decisions: resource types, region selection, quotas, identity model, networking, monitoring, and responsible AI guardrails.
Exam Tip: In case studies, highlight the non-functional requirements first (identity, network, compliance, region availability, quotas). The correct answer is often the one that satisfies the strictest constraint, even if it is not the simplest implementation.
Practice note for Resource planning: regions, quotas, and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Security and identity: keys, managed identity, and RBAC: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Responsible AI governance: safety, privacy, and compliance basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Monitoring and operations: logging, alerts, and reliability patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Domain practice set: Plan and manage (50+ questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Resource planning: regions, quotas, and service selection: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Security and identity: keys, managed identity, and RBAC: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Responsible AI governance: safety, privacy, and compliance basics: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Monitoring and operations: logging, alerts, and reliability patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Domain practice set: Plan and manage (50+ questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-102 frequently tests service selection under constraints. “Azure AI services” (formerly Cognitive Services) covers prebuilt capabilities (Vision, Language, Speech, Translator, Document Intelligence), while Azure OpenAI targets generative AI and embedding scenarios. Azure AI Search often appears as the retrieval layer for RAG (retrieval-augmented generation), and Azure Machine Learning may appear when you need custom model training/MLOps beyond prebuilt offerings.
Resource planning begins with region and availability. Many scenario stems include “must be in region X” or “data residency in country Y.” Not all services (and not all model families/versions in Azure OpenAI) are available in every region. You must also consider quotas (tokens/min, requests/min, or service limits) and throughput requirements. If an app needs bursty traffic, look for patterns like multiple deployments, load balancing across endpoints, or fallbacks to smaller models.
Common trap: Choosing a general LLM to “do OCR” or “parse PDFs” when Document Intelligence is the intended service. The exam expects you to use specialized extractors for reliability and cost, and then optionally use an LLM to summarize or normalize extracted text.
Exam Tip: When you see “must cite sources,” “answers must be grounded,” or “prevent hallucinations,” assume an architecture that includes a retrieval store (often Azure AI Search) + embeddings + prompt instructions to reference retrieved passages.
Provisioning is where exam scenarios shift from “what service” to “how you run it.” You should be comfortable with resource scope (subscription/resource group), naming and environment separation (dev/test/prod), and endpoint management. For Azure AI services, you typically provision a resource instance and use its endpoint + auth method. For Azure OpenAI, you also create model deployments (a deployment is not the same as the resource) and then call the deployment via the resource endpoint.
Quotas are a recurring theme. Azure OpenAI enforces quotas per region and sometimes per model family; scenario questions may describe throttling, 429 errors, or insufficient throughput. Correct operational choices include requesting quota increases, distributing traffic across deployments/regions (if allowed), and implementing client-side retry with backoff. Also watch for “multiple teams” and “least privilege”—that pushes you toward separate resources or separate deployments with scoped access patterns.
Common trap: Confusing the Azure OpenAI resource endpoint with a deployment. On the exam, “create a deployment” is usually the step required to select a model and make it callable, even after the resource exists.
Exam Tip: If the stem mentions “blue/green,” “canary,” or “minimal downtime,” think in terms of multiple deployments (old/new) with controlled traffic shifting, plus monitoring to validate before full cutover.
Security and identity are high-yield on AI-102. You’ll see two broad access patterns: key-based authentication (using access keys) and Microsoft Entra ID (Azure AD) with role-based access control (RBAC). The exam typically prefers Entra ID for production scenarios, especially when “no secrets in code,” “rotate credentials,” or “use managed identity” appears.
Keys are simple and common for quick starts, but they are secrets. If a scenario mentions a CI/CD pipeline, key rotation, or incident response, expect guidance like storing secrets in Azure Key Vault and rotating keys without downtime. Managed identity is the standard pattern for Azure-hosted workloads (App Service, Functions, AKS, Logic Apps) to authenticate to other Azure resources without storing credentials.
Common trap: Treating authentication as authorization. Entra ID gets you an identity; RBAC grants permissions. If the question says “user can sign in but cannot call the API,” the fix is often role assignment, not changing auth type.
Exam Tip: When you see “internal service-to-service calls,” “no credentials stored,” or “automated rotation,” choose managed identity. When you see “external client app,” think carefully: client-side direct calls may require different patterns (e.g., a backend proxy) to avoid exposing keys and to enforce authorization.
Networking questions usually encode a data boundary requirement: “must not use public internet,” “restrict to VNet,” “allow only corporate IP ranges,” or “exfiltration risk.” The controls you should recognize are private endpoints (Private Link), service firewall/IP rules, and network segmentation through VNets and subnets. If the stem is strict about “no public endpoint exposure,” private endpoints are the typical correct direction.
Firewall rules and IP allowlists are helpful but still rely on the public endpoint; they limit who can reach it but do not remove public exposure. Private endpoints place the service behind a private IP in your VNet, enabling private routing and tighter control with NSGs and private DNS. In many architectures, you combine private endpoints for Azure AI Search and storage with VNet-integrated compute (Functions/AKS) so data never exits private routing.
Common trap: Picking “IP restriction” when the requirement is “no public internet.” IP restriction reduces exposure but does not satisfy strict private connectivity language.
Exam Tip: When network questions mention “name resolution fails,” “can’t reach private endpoint,” or “works on-prem but not in VNet,” suspect private DNS misconfiguration (wrong zone, missing link to VNet, or clients not using the correct DNS).
Monitoring and operations show up as “how do you detect failures,” “how do you troubleshoot throttling,” “how do you alert,” and “how do you prevent runaway costs.” Azure Monitor provides metrics and logs; Application Insights adds distributed tracing for app components. For AI services, you often monitor request rates, latency, error codes (401/403/429/5xx), and saturation patterns. For generative AI, you also care about token usage, response times, and safety filter actions as part of operational governance.
Reliability patterns are testable: retries with exponential backoff for transient errors, circuit breakers to protect dependencies, and queue-based load leveling when downstream services throttle. If the scenario mentions spikes or unpredictable load, an async pattern (queue + worker) is often more reliable than scaling a synchronous API alone.
Common trap: Logging prompts and full user content by default. Responsible AI governance overlaps here: privacy and compliance requirements can make verbose logging unacceptable. Prefer redaction, hashing, or logging metadata (token counts, timings, status codes) instead of raw content unless explicitly allowed.
Exam Tip: If the question asks for “reduce cost” in Azure OpenAI scenarios, look for levers like smaller model selection, caching, limiting max tokens, controlling retrieval size (top-k), and using embeddings + retrieval to avoid long prompts.
This section ties the chapter lessons together the way AI-102 does: as multi-constraint scenarios. A typical planning stem includes region constraints, identity requirements, network isolation, and an operational goal (SLA, cost ceiling, or auditability). Your approach should be systematic: (1) pick the service(s), (2) verify region/model availability and quotas, (3) select identity and access model, (4) enforce network boundaries, (5) add monitoring/budget controls, and (6) implement responsible AI guardrails.
Responsible AI governance is often implicit. If you see “user-generated content,” “public-facing chatbot,” “children,” “medical/legal advice,” or “PII,” the exam expects safety measures: content filtering, prompt instructions that refuse unsafe requests, and a design that minimizes data retention. Privacy and compliance basics can drive architecture choices like private endpoints, encryption-at-rest defaults, and data minimization in logs and storage.
Common trap: Solving only the “model” part and ignoring operations. If a scenario asks for “most reliable” or “easiest to manage,” operational features (alerts, budgets, retry policies, and isolation) usually determine the best answer.
Exam Tip: When two answers both “work,” choose the one that reduces secret management, narrows network exposure, and improves diagnosability. AI-102 rewards secure-by-default, operable-by-default designs more than quick-start simplicity.
1. A company is deploying an Azure AI solution to support users in Japan and Australia. The solution must minimize end-user latency and must be deployable only in regions where the required Azure AI service is available. What should you do first?
2. You are building a web app that calls Azure OpenAI. Security requires that no secrets are stored in code or configuration files, and access must be limited to only the web app. What should you use to authenticate from the app to the service?
3. A team needs separate dev, test, and prod environments for an Azure AI solution. They must ensure that developers can manage dev resources but cannot access production data or modify production AI resources. Which approach best meets the requirement?
4. A healthcare company is deploying an AI summarization feature. Compliance requires tracking model outputs for quality review, but the company must minimize retention of sensitive patient data. Which design best aligns with responsible AI governance and privacy requirements?
5. An Azure AI solution is intermittently failing due to spikes in request volume. You need to detect the issue quickly and improve reliability without significantly increasing cost. What should you implement?
This chapter maps to the AI-102 objective area focused on implementing generative AI solutions with Azure OpenAI, including reliable prompting, retrieval-augmented generation (RAG), and safety/evaluation. Expect the exam to test not only feature recognition (what a model/deployment/embedding is) but also decision-making: which capability solves a given constraint (latency, grounding, security, and cost) with the fewest moving parts.
A recurring exam theme is separating platform responsibilities (Azure OpenAI resource, deployments, quotas/limits, monitoring) from application responsibilities (prompt structure, grounding, caching, and evaluation). When a question mixes them, look for the layer that can actually enforce the requirement. For example, “prevent output in a disallowed category” belongs to safety filters and policy, while “reduce token usage” is largely a prompt + chunking + caching problem.
The last lesson in this chapter is a domain practice set. While you won’t see questions here, you should read this chapter like a checklist: can you explain the tradeoff, name the Azure feature that implements it, and identify the common trap answer that sounds right but is incomplete?
Practice note for Azure OpenAI basics: models, deployments, tokens, and limits: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Prompt engineering for reliability: structure, grounding, and constraints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for RAG design: embeddings, vector search, and citation patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Safety and evaluation: content filters, red teaming, and quality checks: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Domain practice set: Generative AI (60+ questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Azure OpenAI basics: models, deployments, tokens, and limits: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Prompt engineering for reliability: structure, grounding, and constraints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for RAG design: embeddings, vector search, and citation patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Safety and evaluation: content filters, red teaming, and quality checks: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Domain practice set: Generative AI (60+ questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
The exam frequently starts with basics: you provision an Azure OpenAI resource, then create deployments that point to specific models. A model is the underlying capability (for example, chat/completions or embeddings); a deployment is the named endpoint your application calls. Many candidates confuse “model name” with “deployment name.” On AI-102, assume your code calls the deployment, and your operations team can change which model sits behind it without rewriting code.
Model selection is usually tested through requirements language. If the prompt asks for “generate answers with citations grounded in documents,” that is an application pattern (RAG) and not a special model—so pick a chat-capable model and design grounding. If the requirement says “semantic similarity, clustering, or vector search,” you need an embeddings model (text-to-vector). If the requirement emphasizes “lowest latency/cost,” the correct answer is often “use the smallest model that meets quality,” plus token controls, not “scale up.”
Tokens and limits appear as operational constraints: context window (how much you can send in), max output tokens (how much you can receive), and throughput constraints (rate limits). The exam often frames this as: “The prompt fails intermittently with a length error” or “responses are truncated.” Correct fixes are to reduce input tokens (shorter prompt, fewer retrieved chunks, summarize history) and/or increase allowed output tokens, not “change to embeddings” or “add more instances.”
In scenario questions, identify whether the issue is (1) model capability mismatch (chat vs embeddings), (2) context window overflow, or (3) quota/rate limiting. These map cleanly to the right remediation.
Prompt engineering on AI-102 is less about “clever wording” and more about reliability: consistent format, constrained scope, and explicit grounding instructions. The exam expects you to know that the system message sets high-priority behavior (role, boundaries, safety constraints), while user messages carry the task and data. When a question asks how to enforce global behavior across a conversation (tone, refusal rules, output schema), the best answer usually involves the system role.
Few-shot prompting is tested as a way to reduce variance and stabilize formatting. If the requirement says “the output must always include these fields” or “classify into one of N labels,” few-shot examples plus explicit constraints (e.g., “choose exactly one label”) are stronger than adding more instructions alone. However, few-shot increases tokens—so it can worsen cost/latency and context window pressure. Expect tradeoff questions.
Structured output is a major reliability theme. The exam will often include “downstream parser fails” or “JSON sometimes invalid.” The correct approach is to constrain output with a clear schema and strict instructions (and, where supported, structured output / JSON mode patterns). In prompts, put the schema near the end, require double quotes, forbid trailing commentary, and specify error handling (e.g., return an object with an error field). Avoid mixing natural language and JSON in the same response unless explicitly required.
To identify correct answers, look for options that: (1) separate instructions from data, (2) specify constraints and format, and (3) explicitly define what to do when information is missing (refusal/unknown). That combination is what the exam rewards.
Embeddings convert text (and sometimes other modalities) into vectors that preserve semantic similarity. On AI-102, embeddings are the foundation for vector search in RAG. The exam will test whether you know when to use embeddings: semantic similarity, nearest-neighbor retrieval, deduplication, clustering, and “find related content” features. If a question is about “searching by meaning” rather than keyword, embeddings are the intended solution.
Vectorization strategy is a practical design point the exam likes. You typically embed documents (or chunks) offline during indexing, then embed the user query at runtime and perform similarity search. The trap is embedding everything at runtime (too slow/costly) or embedding entire large documents as one vector (retrieval becomes imprecise). Another common test concept: maintain consistent preprocessing. If you normalize whitespace, remove boilerplate, or add metadata to the chunk text, do it the same way at embedding time and retrieval time; otherwise similarity can degrade.
Metadata matters. In production RAG, you rarely rely on vectors alone—filters (category, tenant, ACL, language, date) narrow candidates before ranking. The exam may describe multi-tenant or role-based access requirements; the correct retrieval design uses metadata filtering to enforce authorization before the model ever sees the text.
When choosing the “best next step” on the exam, align embeddings choices with the goal: precision (better chunking + metadata filters), recall (broader retrieval, query expansion), or compliance (strict filtering and tenant isolation).
RAG questions typically provide a symptom: hallucinations, missing details, or answers without traceability. Your job is to pick the implementation lever: chunking, ranking, citation pattern, or caching. Chunking is the most-tested. Smaller chunks increase retrieval precision but can lose context; larger chunks preserve context but dilute similarity and waste tokens. A good exam answer mentions balancing chunk size with the model’s context window and the expected question type (fact lookup vs multi-paragraph synthesis).
Ranking and re-ranking show up when “retrieved passages are irrelevant.” The exam expects you to know that first-stage retrieval is approximate; improving results can involve hybrid retrieval, adjusting top-K, adding metadata filters, or using a re-ranker to reorder candidates. Another practical lever: include the user’s question when generating the final answer and explicitly instruct the model to cite only from the provided passages.
Citation patterns are a reliability and governance requirement. A common requirement is “include sources for each claim.” The best pattern: attach source IDs to each chunk, pass them along with the retrieved text, and require the model to reference those IDs inline (e.g., [doc3], [policy-12]) and/or provide a sources list. The trap is asking the model to “add citations” without providing stable identifiers—this invites fabricated references.
Caching is tested as a cost/latency solution. Cache embeddings for repeated queries, cache retrieval results for stable corpora, and cache final answers only when acceptable (be careful with personalization and authorization). If the stem mentions “same question asked frequently” and “high cost,” caching is often the most direct fix.
On AI-102, strong answers clearly separate steps: index & embed content → retrieve & rank → construct grounded prompt → generate with citation constraints → log/evaluate.
Safety is not optional on the exam. You will be tested on layered defenses: platform content filters, application-level policies, and prompt-level guardrails. Content filtering typically covers both prompts and completions, helping detect categories of harmful content. The exam commonly asks what to do when a solution must “block disallowed content” or “reduce risk of unsafe outputs.” The correct answer is rarely “just add a warning in the prompt.” You combine filtering with refusal behavior and logging for audit.
Jailbreak resistance is usually framed as “users try to override system instructions” or “prompt injection from retrieved documents.” Your mitigation is defense-in-depth: keep critical rules in the system message, sanitize and delimit retrieved content (e.g., clearly mark it as untrusted quoted context), and instruct the model to treat retrieved text as data, not instructions. For prompt injection, retrieval filtering helps too—exclude untrusted sources, apply allow-lists, and use metadata-based trust scoring.
Red teaming and evaluation are part of safety and quality. The exam may describe a need to “test harmful scenarios before release.” Look for answers that include adversarial testing, dataset-based evaluation, and continuous monitoring rather than a one-time manual review. Quality checks can include factuality against sources, refusal correctness, and structured output validity.
In answer choice elimination, remove options that rely on a single control. AI-102 prefers layered mitigations that address both user input and model output paths.
Troubleshooting questions are where AI-102 blends engineering and operations. Latency problems usually come from: large prompts (too many tokens), slow retrieval (unoptimized vector search), or repeated calls (no caching). The best fixes are often “reduce tokens” (shorter instructions, fewer chunks, summarize chat history), “optimize retrieval” (top-K tuning, filters, hybrid search), and “cache appropriately.” Switching to a larger model rarely reduces latency; it often increases it.
Hallucinations are typically a grounding failure. Diagnose by asking: did retrieval return the right passages, and did the prompt force the model to use them? Improvements include: better chunking, re-ranking, stricter instructions (“answer only from sources; otherwise say ‘I don’t know’”), and citation enforcement. If the exam stem says “the model invents citations,” your fix is to pass explicit source IDs and forbid unstated sources, not to “raise temperature” (which worsens it).
Cost issues are token issues plus throughput. Look for cost controls: limit max output tokens, reduce retrieved context, compress/systematize prompts, use smaller models when acceptable, and cache embeddings and retrieval results. Also watch for hidden cost multipliers: few-shot examples repeated every turn, verbose system messages, and passing entire documents instead of chunks.
Finally, expect scenarios that require balancing all three: latency, hallucinations, and cost. The exam’s preferred path is systematic: measure tokens and retrieval quality first, then adjust chunking/ranking/caching, and only then consider model changes.
1. You are building a customer support chatbot with Azure OpenAI. During load testing, you see frequent responses ending mid-sentence and the API returns errors indicating a token limit was reached. You want to reduce the chance of truncation without changing the model. What should you do first?
2. A company needs an internal Q&A assistant that must answer only from an approved policy document set and include citations to the source passages. The documents change weekly. Which design best meets the requirement with the fewest moving parts?
3. You are implementing RAG. Users complain that the assistant sometimes cites irrelevant snippets. You want to improve retrieval quality before changing the language model. Which change is most likely to help?
4. A healthcare organization must prevent the assistant from generating disallowed content categories (for example, hate or self-harm) regardless of how users prompt it. Which control is the most appropriate to enforce this requirement?
5. You are asked to evaluate a new prompt and RAG configuration before production. The goal is to uncover jailbreaks and ensure answers are grounded in retrieved sources. What is the best approach?
AI-102 increasingly tests whether you can move from “a good chat prompt” to an agentic application: a model that can select tools, follow an orchestration pattern, maintain conversation state, and ground outputs in enterprise data safely. In exam terms, you’re expected to recognize design choices that reduce hallucinations, limit actions to approved boundaries, and keep user data protected while still enabling multi-step task completion.
This chapter maps to the “Implement an agentic solution” outcomes: tool use, orchestration, grounding, and conversation state. Expect scenario questions that describe a bot, a workflow, or an incident (wrong tool call, prompt injection, missing citations, runaway costs) and ask what you should change. The correct answers usually emphasize: explicit tool contracts, least-privilege boundaries, deterministic validation, state minimization, and measurable guardrails.
As you read, focus on how Azure AI solutions are assembled: a model endpoint (Azure OpenAI), tool endpoints (APIs, functions, search, databases), and orchestration code. The exam rarely rewards “magic prompting” alone; it rewards architectures that are testable and safe in production.
Practice note for Agent fundamentals: tools, functions, and orchestration patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for State and memory: session context, grounding data, and constraints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Multi-step reasoning and tool use: planning, calling, and validation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Production readiness: guardrails, monitoring, and fallback strategies: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Domain practice set: Agentic solutions (40+ questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Agent fundamentals: tools, functions, and orchestration patterns: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for State and memory: session context, grounding data, and constraints: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Multi-step reasoning and tool use: planning, calling, and validation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Production readiness: guardrails, monitoring, and fallback strategies: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Domain practice set: Agentic solutions (40+ questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Agent fundamentals begin with crisp goals and boundaries. On AI-102, “agent” does not mean autonomous without limits; it means a model-driven loop that can decide when to call tools, while your application constrains what “done” means and what actions are allowed. Good designs separate: (1) the user’s request, (2) the system/developer instructions that define policy and style, and (3) tool permissions that define what the agent can actually do.
Expect exam scenarios where the agent performs an unsafe action (emails the wrong person, deletes data, reveals secrets). The best answer usually isn’t “tell the model to be careful,” but “restrict the tool boundary.” Examples: expose a read-only customer lookup tool instead of direct database access, require a confirmation step before executing a write tool, and enforce tenant scoping server-side (never rely on the model to enforce access rules).
Exam Tip: If a question mentions “prevent unauthorized actions” or “ensure the agent can’t exceed scope,” look for answers involving least privilege (RBAC/managed identity), allowlisted tools, and server-side authorization checks—not just prompt instructions.
Common trap: mixing policy into user-visible prompts. System/developer instructions are the correct place for non-negotiable rules, while user prompts should remain task-focused. Another trap: assuming the model will “remember” constraints; in practice, constraints must be enforced in the orchestrator and tool layer.
Tool (function) calling turns free-form text into structured actions. The exam targets whether you can design tool schemas that are explicit, verifiable, and resilient to model mistakes. A strong tool design uses strict parameter types, required fields, enumerations, and clear descriptions that reduce ambiguity (for example, ISO-8601 dates, explicit currency codes, and bounded integer ranges).
Validation must happen outside the model. If the model produces malformed JSON, an out-of-range value, or a parameter that violates business rules, your orchestrator should reject it and either retry with a corrected instruction or fall back to a safe alternative. Retries should be bounded and observable; “infinite retry until it works” is a production anti-pattern and a frequent exam distractor.
Exam Tip: When asked how to “increase reliability of tool calling,” choose server-side schema validation + deterministic error handling. Answers that only say “improve the prompt” are typically incomplete unless paired with validation.
"priority": "low|medium|high") over free text.Common trap: letting the model call arbitrary URLs or execute arbitrary code. In AI-102 context, the correct architecture is an allowlist of tools with explicit schemas, with the orchestrator mediating calls. Another trap is ignoring partial failures in multi-tool workflows; correct solutions include step-level validation and compensation logic for failed steps.
State and memory are core to agentic apps: the model needs enough context to act consistently, but the system must minimize sensitive data exposure and control cost. AI-102 scenarios often ask what to store, for how long, and where. A practical pattern is to separate (1) short-term session state (recent turns, tool results), (2) long-term user preferences (opt-in, curated), and (3) enterprise knowledge (retrieved on demand via search/RAG rather than stored in chat history).
Data minimization is a key production readiness theme. Do not “stuff everything” into the prompt: it raises token cost, increases leakage risk, and can degrade relevance. Instead, summarize older turns, store structured state (like selected product ID, date range), and retrieve grounding content only when needed. Apply privacy controls: redact or avoid storing PII, apply retention policies, and use encryption and access controls for stored transcripts.
Exam Tip: If a question mentions privacy, compliance, or “least data,” the best answer usually involves minimizing prompt contents, using retrieval instead of long chat history, and applying retention/redaction—not adding more context.
Common trap: using long-term memory as a dumping ground for every user message. On the exam, persistent memory should be deliberate, user-consented, and auditable. Another trap is assuming “grounding data” must be permanently stored; often it should be retrieved transiently per request and discarded or logged only in redacted form.
Orchestration is how you structure multi-step reasoning and tool use. AI-102 tends to test whether you can choose an appropriate pattern given requirements for reliability, latency, and safety. Three common patterns appear in scenarios:
On the exam, the “right” orchestration pattern is often the one that reduces risk. For example, if the scenario includes sensitive actions (refunds, account changes), a router plus deterministic workflow (human approval, explicit confirmation) is often safer than a free-form planner that might attempt unexpected operations.
Exam Tip: When you see “must be auditable” or “must follow a fixed process,” prefer orchestrations that externalize the process into code/workflow steps with logs, rather than letting the model improvise the whole procedure.
Common trap: assuming multi-agent is automatically better. It can improve accuracy, but it can also multiply token usage and failure points. Look for constraints like “low latency” or “cost sensitive”—those often imply a simpler router or a constrained planner-executor with bounded steps and caching of tool results.
Grounding is how you keep the agent’s outputs anchored to trusted sources. In AI-102, you’re expected to combine Retrieval-Augmented Generation (RAG) with tool use: Azure AI Search (or equivalent retrieval) provides relevant passages, while tools provide authoritative, real-time facts (ticket status, inventory counts, policy versions). The model then answers using only the supplied context and includes citations when required.
Questions commonly test whether you can improve answer correctness. Typical fixes include: improving chunking and indexing strategy, using semantic ranking/hybrid search, filtering by metadata (tenant, document type, security labels), and requiring citations. If the scenario mentions “users complain about fabricated policy statements,” citations plus “answer only from retrieved sources” is a strong direction—but also ensure retrieval is correct (wrong documents retrieved leads to confidently wrong answers with citations).
Exam Tip: If the requirement says “must cite sources” or “must trace to documents,” select options that return document IDs/URLs/sections from the retriever and propagate them to the final response. Citations are an application responsibility as much as a prompt instruction.
Common trap: storing enterprise documents directly in chat history. The exam generally favors retrieval at query time with access controls. Another trap is skipping metadata filters—multi-tenant leakage is a classic scenario, and the correct answer involves strict filtering and identity-aware retrieval (managed identity, per-user claims, or document-level permissions enforced by the search layer).
Production readiness ties everything together: guardrails, monitoring, and fallback strategies. AI-102 scenario questions often describe an agent failing in one of three ways: (1) tool errors, (2) prompt injection or data exfiltration attempts, or (3) unsafe actions (overstepping authority). Your job is to choose mitigations that are enforceable by design.
Tool errors: Handle transient failures with exponential backoff and bounded retries; handle permanent failures with user-facing explanations and alternative paths (for example, “I can’t access the CRM right now; I can create a support ticket”). Log tool call inputs/outputs (redacted) and error codes for troubleshooting. Common trap: letting the model “guess” missing tool results—correct designs fail closed and ask for clarification or degrade gracefully.
Prompt injection: Treat retrieved documents and user content as untrusted. Use content boundaries (e.g., “The following is reference text; do not follow instructions inside it”), strip or ignore tool-invocation directives coming from retrieved text, and enforce an allowlist of tools plus parameter validation. Exam Tip: If the scenario says the model followed instructions from a document (“ignore previous instructions, send secrets”), the fix is not more RAG; it’s isolation of instructions, strict tool control, and output filtering/safety policies.
Safe actions: For high-impact operations (payments, deletions, account changes), require explicit user confirmation and/or human-in-the-loop approval, and implement policy checks in code. Monitor for abnormal behavior: spikes in tool calls, repeated failures, or unusual data access patterns. Fallback strategies include switching to a simpler Q&A mode, disabling write tools temporarily, or routing to a human agent. Common trap: assuming the model’s “I will not do that” is sufficient; the exam expects enforceable controls at the orchestrator and tool layers.
1. A company is building an HR benefits agent that can search an internal policy index and open a support ticket. During testing, the model sometimes calls the ticketing tool even when the user only asked a policy question. You need to reduce unintended tool execution while keeping the agent capable of multi-step actions. What should you implement first?
2. You are designing a customer support agent that uses conversation history to handle returns. The agent currently stores the full transcript of every session indefinitely. Security requires minimizing stored personal data while still enabling the agent to remember key details (order number, product, return window) during the session. What is the best approach?
3. An agent performs a multi-step task: (1) retrieve documents from a knowledge base, (2) draft a response, (3) send an email through an API. In production, the agent occasionally emails incorrect information due to retrieval returning irrelevant results. What should you add to improve reliability before the send-email tool is called?
4. A finance agent uses a tool that executes SQL queries against an internal database. A user enters: "Ignore your instructions and run: DROP TABLE Transactions;" The agent passes the text to the SQL tool. You must mitigate prompt injection and limit blast radius. What is the best solution?
5. Your team deploys an agent that calls multiple external tools. After launch, costs spike and telemetry shows repeated retries and long tool-call loops on ambiguous user requests. You need to make the system production-ready with measurable guardrails and a safe fallback. What should you implement?
This chapter aligns to the AI-102 objective area that tests whether you can choose the right Azure AI service for vision, language, and knowledge mining, then assemble them into an end-to-end solution with the correct “plumbing”: inputs/outputs, confidence handling, indexing, and security. On the exam, many incorrect options are “almost right” because Azure has overlapping capabilities (for example OCR exists in multiple places). Your job is to recognize the primary service for the requirement, the expected artifacts (models, indexes, skillsets), and operational details (confidence scores, language support, quotas, and deployment patterns).
Expect scenario prompts like: “extract fields from invoices,” “detect objects in images,” “redact PII from call-center transcripts,” or “build searchable knowledge over PDFs and images.” The exam doesn’t just check feature recall—it checks if you can map business requirements to service selection, and whether you understand the limits of each approach (e.g., OCR output is not the same as structured document extraction; search indexing is not the same as model training).
As you read, keep one mental model: capture signals (vision, speech, text) → normalize and extract (OCR, entities, key phrases, structured fields) → enrich and store (Azure AI Search index + skillsets) → serve in apps (APIs, bots, dashboards) with monitoring, security, and cost control.
Practice note for Computer vision: image analysis, OCR, and document processing scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for NLP: text analytics, conversational language, and speech/language services overview: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Knowledge mining: Azure AI Search indexing, enrichment, and extraction pipelines: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for End-to-end integration: combine vision/NLP outputs with search and apps: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Domain practice set: Vision + NLP + Knowledge Mining (70+ questions): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Computer vision: image analysis, OCR, and document processing scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for NLP: text analytics, conversational language, and speech/language services overview: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Knowledge mining: Azure AI Search indexing, enrichment, and extraction pipelines: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for End-to-end integration: combine vision/NLP outputs with search and apps: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
AI-102 frequently tests whether you can select the correct vision capability based on the output needed: descriptive analysis, text extraction, or domain-specific classification/detection. For general-purpose image understanding (tags, captions, object/people detection, background removal depending on capability set), you’re typically in the “image analysis” family. If the requirement is “read text in an image,” you pivot to OCR. If the requirement is “identify our company’s product SKUs” or “detect a defect type unique to our factory,” you likely need a custom model.
Exam Tip: Look for wording like “read,” “extract text,” “printed/handwritten,” “receipt,” or “signage.” Those keywords usually indicate OCR (or Document Intelligence if the next step is structured fields). If the wording says “classify into custom categories,” “detect custom objects,” “train with labeled images,” it signals a custom vision approach rather than off-the-shelf analysis.
Common trap: choosing OCR when the prompt wants structured extraction. OCR returns lines/words with bounding boxes; it does not inherently understand “InvoiceNumber” or “TotalDue.” Another trap is assuming generic image analysis can recognize your proprietary label types or niche parts—it may not, so custom training is required.
How to identify the correct answer: isolate the primary deliverable (labels vs text vs domain-specific inference), then check constraints (offline vs cloud, latency, language/format support). On AI-102, “service selection” questions often hide one decisive constraint—e.g., “handwritten notes in scanned forms” pushes you beyond simple image tags into OCR/Document Intelligence workflows.
Document processing scenarios are a major exam theme because they combine OCR with structure. The exam expects you to understand that Document Intelligence (historically called Form Recognizer) is designed to extract fields, tables, and key-value pairs from documents like invoices, receipts, IDs, contracts, and custom forms. Where OCR gives you text blocks, Document Intelligence gives you semantically meaningful outputs, often already mapped to business concepts (for prebuilt models) or mapped to your schema (for custom models).
Know the workflow: ingest a document (PDF/image) → run a prebuilt or custom model → parse returned fields, table cells, and bounding regions → apply business rules with confidence thresholds → store results (database or search index) and optionally route low-confidence items for human review.
Exam Tip: Confidence scores are not “accuracy guarantees.” On the exam, a correct design often includes a threshold (e.g., if confidence < X, send to human validation). A common trap is choosing “reject all low confidence” rather than “route to review,” especially for compliance and finance scenarios.
Be ready for details about tables vs fields. Tables often return row/column structures plus cell content and bounding boxes; fields return name/value pairs and types (string, date, number) depending on model. A scenario mentioning “line items” on invoices (quantity, unit price, description) is a strong signal you must handle tables, not just key-value pairs.
Also watch for the trap of treating multi-page PDFs as single images; robust designs handle page ranges and return per-page results. If the scenario includes “store bounding boxes for highlighting in a UI,” ensure your chosen service exposes geometry/regions in the response—Document Intelligence and OCR both do, but the structured output differs.
In AI-102, “NLP selection” questions often test whether you can distinguish common text analytics tasks and choose the appropriate capability: sentiment analysis, key phrase extraction, named entity recognition (NER), and PII detection/redaction. The exam commonly frames these as customer feedback processing, support ticket triage, compliance monitoring, or summarizing large volumes of text prior to search indexing.
Map requirements to outputs: sentiment yields polarity/score for opinions; key phrases yields topical phrases for tagging; entities yields structured mentions like people, organizations, locations, dates; PII yields detection of sensitive types (e.g., phone numbers, addresses, IDs) and often supports masking/redaction patterns in your pipeline.
Exam Tip: If the scenario says “remove personal data before storing,” your answer should include PII detection and a data-handling step (masking/redaction) prior to persistence or indexing. A common trap is to index first and redact later—this violates “data minimization” requirements and is often presented as an incorrect option.
Another frequent trap is mixing “classification” with “sentiment.” Sentiment is opinion polarity; classification is assigning a category label such as “billing issue” vs “technical issue.” If the question says “route tickets by topic,” you need classification (custom or predefined), not sentiment alone—even if sentiment is also useful for prioritization.
How to pick the correct answer under pressure: underline the noun in the prompt (sentiment, entities, phrases, PII) and match it to the output artifact expected in the application (score, entity list, tags, redacted text). Then validate that the design accounts for multilingual needs and downstream storage, which are classic exam constraints.
Conversational scenarios in AI-102 test your understanding of how users interact with systems via natural language and how you extract intent and entities to drive actions. Exam prompts may mention a “chatbot,” “virtual agent,” “call center automation,” or “multi-turn conversation.” The key is recognizing the building blocks: a client channel (web/mobile/Teams), bot orchestration, language understanding (intent + entities), and state management for multi-turn dialogues.
Core concepts to know: intent (what the user wants), entities (parameters like date, product, location), and conversation state (what’s already been collected). If a scenario says “user can say ‘book it for next Friday’ after selecting a destination,” you must preserve state so the second utterance can be resolved in context.
Exam Tip: Multi-turn requirements are often the differentiator. If the prompt includes follow-up questions, ambiguity resolution, or confirmation (“Did you mean…?”), choose an approach that supports dialog flow and state, not just one-shot intent detection.
Common trap: treating a bot as only “Q&A.” Some solutions are retrieval-based (FAQ over documents), but if the bot must trigger actions (create ticket, schedule appointment), you need reliable intent routing and parameter extraction. Another trap is ignoring language variation—synonyms and user phrasing require robust training data or a model that generalizes well; the exam may hint this with “users phrase requests in many ways.”
Remember that conversational solutions often combine with knowledge mining: a bot answers by querying an Azure AI Search index enriched from documents, and it can also log transcripts for analytics (after applying PII redaction when required).
Knowledge mining is a high-yield AI-102 area because it forces you to connect extraction (vision/NLP) to search experiences. Azure AI Search is the hub: you define an index (fields, analyzers, filterable/facetable/sortable attributes), populate it via an indexer (pulling from data sources like Blob Storage), and optionally apply an enrichment pipeline using skillsets (OCR, entity extraction, key phrases, language detection, and custom skills).
The exam checks if you understand why skillsets exist: to transform raw content into searchable structure. A PDF in blob storage is not immediately useful unless you extract text, normalize language, split content into chunks, and project results into index fields. Index design matters—if the scenario requires filtering by “customerId” or “documentType,” those must be separate filterable fields, not buried in a single content blob.
Exam Tip: When the question mentions “make scanned PDFs searchable,” think: Blob Storage data source + indexer + skillset with OCR + index projection. A common trap is selecting “store in Cosmos DB and query it” as a substitute for full-text search and faceting; the exam expects Azure AI Search for search-centric requirements.
Also note enrichment outputs: entities can become facets; key phrases can become tags; OCR text becomes the primary searchable content; document metadata (path, lastModified) supports sorting and filtering. If security is hinted (per-user access), design for filtering results by security metadata (e.g., ACL fields) and enforce it at query time.
Finally, watch the “reindexing” trap: changes to skillsets or index schema often require re-running indexing. On the exam, a correct operational approach includes a plan for incremental indexing, monitoring indexer runs, and managing cost when processing large document sets.
This section is about the integration patterns the exam loves: combining vision/NLP outputs with Azure AI Search and then exposing it to an application (web app, internal portal, bot). The key coaching tactic: translate the scenario into a pipeline diagram in your head, then verify each component produces the required artifact.
Example mapping approach (without turning it into a quiz): if the scenario includes “photos of shelves” plus “find products and make them searchable,” start with image analysis or custom object detection (if product set is specific), then store extracted labels/IDs as structured fields. If it includes “scanned contracts” plus “search by party name and effective date,” start with OCR/Document Intelligence for text + fields, then apply NLP entity extraction for parties/dates, then index all outputs in Azure AI Search with facetable fields for those entities.
Exam Tip: The fastest way to eliminate wrong answers is to check whether the proposed service can output the structure the app needs. Search experiences need fields; dashboards need normalized values; compliance needs redaction logs. If an option only returns unstructured text, it’s often insufficient even if it “works” in a demo.
Common integration traps: (1) forgetting language detection/translation when content is multilingual; (2) indexing sensitive data before PII redaction; (3) assuming OCR alone provides semantic fields; (4) storing enriched JSON but not projecting it into the search index, making it unqueryable. Another trap is ignoring latency: real-time app flows may call vision/NLP on-demand, while knowledge mining usually uses batch indexing—exam scenarios often imply which you need via words like “near real-time” vs “nightly ingestion.”
As you move into the domain practice set for Vision + NLP + Knowledge Mining, keep mapping each requirement to (a) the correct extraction service, (b) the artifact it produces (fields/entities/phrases), and (c) how that artifact becomes searchable, filterable, and secure in Azure AI Search. That mapping skill is what AI-102 most consistently rewards.
1. A company needs to extract the invoice number, vendor name, invoice date, and total amount from scanned PDF invoices. The layout varies by vendor, and the solution must return structured fields (not just raw OCR text) with confidence scores. Which Azure AI service should you use?
2. You are building a searchable knowledge base over thousands of PDFs and images stored in Azure Blob Storage. Requirements: extract text from images, detect key phrases and entities from the extracted text, and make the enriched content searchable in one index. Which approach best meets the requirement with the correct Azure AI Search artifacts?
3. A call center wants to redact personally identifiable information (PII) such as phone numbers, email addresses, and credit card numbers from chat transcripts before storing them. Which Azure AI service should you use to detect and redact PII in text?
4. You are implementing an app that lets users upload product photos. The app must identify objects in the image (for example, 'backpack', 'shoe') and return tags with confidence values, without training a custom model. Which service should you use?
5. You built an Azure AI Search index that is enriched using a skillset (OCR + entity recognition). A new requirement states that some fields are sensitive and must not be returned to clients, but the fields must remain searchable for internal users. What is the best way to meet this requirement?
This chapter is where you convert “I studied the objectives” into “I can pass the exam under exam conditions.” AI-102 rewards candidates who can recognize which Azure AI capability fits a scenario, pick the right configuration details, and avoid subtle constraints hidden in wording. Your final prep should therefore be less about reading and more about performing: taking a full mock exam in two parts, analyzing weak spots, and running an objective-by-objective rapid review until your choices become automatic.
Across the lessons in this chapter—Mock Exam Part 1, Mock Exam Part 2, Weak Spot Analysis, Exam Day Checklist, and the Final objective-by-objective rapid review—you will practice the same skills the real exam measures: planning and managing an Azure AI solution; implementing generative AI with Azure OpenAI (prompting, RAG, safety, evaluation); implementing agentic patterns (tool use, orchestration, grounding, conversation state); building computer vision and document intelligence solutions; deploying NLP capabilities; and implementing knowledge mining with Azure AI Search. The goal is not to “know everything,” but to consistently identify the highest-likelihood correct answer by mapping scenario signals to the right service, feature, and security/deployment constraint.
Exam Tip: Treat your mock exam as a diagnostic instrument, not a confidence test. The value is in the post-review: the error log, remediation, and retake strategy.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Final objective-by-objective rapid review: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Practice note for Final objective-by-objective rapid review: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.
Run your full mock exam in two timed blocks (Mock Exam Part 1 and Part 2) to mirror real-world cognitive fatigue. Set a strict environment: one screen, no notes, no pausing, no searching, and no “just checking” documentation. Your brain needs to rehearse recall and elimination under time pressure, not recognition with references.
Use a pacing plan: on the first pass, answer what you know, mark what you don’t, and move on. Your objective is to complete the first pass with time left for review. Many candidates lose points by over-investing early, then rushing late—especially on scenario-heavy items that involve Azure AI Search indexing pipelines, managed identities, network isolation, or RAG grounding steps.
Exam Tip: When you feel stuck, don’t “solve harder.” Instead, identify what domain it belongs to (management/security vs. generative AI vs. search vs. vision vs. language) and eliminate options that clearly belong to a different domain. AI-102 questions are often designed so two answers sound plausible, but only one matches the stated constraint (for example: private networking, data residency, or evaluation requirement).
Simulate interruption resistance: no music, no phone, and practice a short reset routine after each block (stand, water, 2 minutes breathing) to imitate the transition between Part 1 and Part 2.
Before reviewing results, map every missed or uncertain item back to an objective domain. This prevents “random studying” and forces remediation where it matters. A strong AI-102 mock exam includes a balanced spread of: (1) planning and managing Azure AI resources; (2) Azure OpenAI generative patterns including RAG, safety, and evaluation; (3) agentic solutions with tool orchestration and state; (4) computer vision and document intelligence; (5) NLP features like text analytics and conversational services; and (6) knowledge mining with Azure AI Search indexing and enrichment.
Use a simple coverage grid and tag each item with the primary objective and a secondary one. Example: a scenario about RAG with citations might be “Azure OpenAI + Azure AI Search,” while a scenario about secure access to models might be “resource security + networking + managed identity.” This is exactly how the exam blends skills: few questions live in a single silo.
Exam Tip: If your mock score is “good but inconsistent,” it usually means your coverage is skewed. Candidates often over-study generative AI prompts and under-study resource governance and Azure AI Search pipelines—yet those are frequent differentiators in real exams.
Weak Spot Analysis is not “reviewing wrong answers.” It is identifying the reason you chose the wrong option and turning that reason into a fixable pattern. Build an error log with four fields: objective tag, what you chose, why it felt right, and the rule that makes it wrong. Then add the “correct rule” you will apply next time.
Common root causes tend to cluster: misreading constraints (private network, PII handling, latency), confusing similar services (OCR vs document intelligence vs vision), or selecting the right service but wrong configuration (authentication method, indexer vs skillset, vector search prerequisites, content safety placement). Your remediation should match the root cause: if it’s a reading error, practice constraint extraction; if it’s a service confusion, write a one-paragraph differentiator; if it’s configuration, build a checklist.
Exam Tip: Pay special attention to “almost correct” answers. AI-102 often tests the last mile: managed identity vs keys, private endpoint vs firewall rule, index schema vs skillset, or evaluation vs content filtering. Your error log should capture those precise distinctions.
Finish your review by running a mini rapid review: for each missed objective, state the default service, the key differentiator, and the most common constraint that changes the answer.
AI-102 uses traps that punish guessing and reward disciplined reading. Multi-select items are a top trap: candidates pick the “best” single option even when the prompt requires multiple controls (for example, retrieval plus safety plus evaluation). Always match the number of required selections and validate that each selected option independently satisfies a stated requirement.
Case studies hide constraints across tabs and paragraphs. A single line like “must use customer-managed keys,” “must not store prompts,” “requires private connectivity,” or “global rollout with regional failover” can invalidate half the options. Train yourself to extract constraints first, then look for answers.
Exam Tip: When two answers sound plausible, look for an explicit constraint that only one satisfies. If none exists, the exam is likely testing “default best practice” (for example: managed identity, least privilege, separation of duties, or structured evaluation rather than anecdotal testing).
Finally, watch for “feature drift” assumptions. Don’t assume every capability exists in every service tier/region. If the question emphasizes compliance, residency, or isolation, prioritize architecture choices that control network and identity boundaries.
Your final week is about stability. Retake the full mock exam after remediation, but do not retake immediately. Use spaced repetition: remediate today, retest in 48–72 hours, then again near the end of the week. This spacing forces retrieval, which is what you need on exam day.
Use confidence checks per objective rather than raw scores. For each course outcome, ask: can you (a) choose the correct service, (b) justify configuration/security choices, and (c) explain why alternatives fail? If you can’t do all three, you are not exam-ready for that objective even if you sometimes guess correctly.
Exam Tip: If your weakness is broad (many objectives), focus on high-leverage domains: Azure AI Search knowledge mining pipelines, Azure OpenAI RAG architecture and safety, and resource governance (identity/network). Those areas often create the biggest score swings because they combine multiple objectives in one scenario.
Stop heavy studying the night before. Your priority is recall speed and attention control, not new information.
Your Exam Day Checklist should be operational, not motivational. Confirm your ID requirements, appointment time zone, and whether you are taking the exam in a test center or via online proctoring. For online delivery, prepare the room: clean desk, stable internet, power cable, no secondary devices, and a plan for notifications (disable them). Many candidates lose time and composure due to avoidable setup issues.
During the exam, manage attention like a resource. Read the question stem first, then constraints, then options. Use a consistent process: identify the domain, extract constraints, eliminate mismatched services, then decide. If you hit a hard question, mark it and move on—protect your pacing plan.
Exam Tip: When stressed, candidates overcomplicate. Default back to best practices: least privilege, managed identity, private networking when required, and clear separation of ingestion/enrichment/retrieval for knowledge mining. If an option violates these without a stated reason, it’s likely wrong.
After finishing, do a final review of marked items only. Re-reading everything increases the chance you talk yourself out of correct answers. End the exam the same way you practiced: controlled, systematic, and aligned to objectives.
1. A retail company is building a chatbot that answers questions using internal policy documents stored in SharePoint and PDFs. The chatbot must (1) ground answers in the documents, (2) cite sources, and (3) avoid returning content that violates company safety rules. Which approach best meets the requirements with Azure services used in AI-102 scenarios?
2. You are reviewing missed mock exam questions and notice a repeated mistake: selecting the correct Azure AI service, but choosing the wrong security configuration. A solution uses Azure OpenAI and must ensure only a private network can access the endpoint. Which configuration should you validate first during your weak-spot remediation?
3. A company wants an agent-style assistant that can (1) call internal tools (e.g., check inventory, create a support ticket), (2) maintain conversation state across turns, and (3) minimize hallucinations by grounding tool outputs into responses. Which design best fits agentic patterns expected in AI-102?
4. A financial services firm must process scanned loan applications and extract fields such as applicant name, address, loan amount, and employer. The forms vary slightly by region but follow the same general template. Which Azure service is most appropriate to implement this requirement?
5. During your final objective-by-objective review, you encounter a scenario: An enterprise wants to enable employees to search across product manuals, support tickets, and FAQs with semantic ranking and filterable metadata (product, version, region). The solution must return the most relevant passages and allow citations. Which Azure capability should you prioritize?