
Simulate the real exam experience with 50 questions and a 45-minute time limit. Practice with AI-verified answers and detailed explanations.
AI-Powered
Every answer is cross-verified by 3 leading AI models to ensure maximum accuracy. Get detailed per-option explanations and in-depth question analysis.
DRAG DROP - Match the types of machine learning to the appropriate scenarios. To answer, drag the appropriate machine learning type from the column on the left to its scenario on the right. Each machine learning type may be used once, more than once, or not at all. NOTE: Each correct selection is worth one point. Select and Place:
Predict how many minutes late a flight will arrive based on the amount of snowfall at an airport. ______
Predicting “how many minutes late” is predicting a continuous numeric value (minutes). That is the defining characteristic of a regression problem: the model learns a relationship between input features (for example, snowfall amount, airport, time of year, historical delay patterns) and a numeric target. Why not classification? Classification would apply if the output were a category such as “on time / late” or “late by 0–15, 16–30, 31+ minutes.” Those are discrete classes, not a continuous number. Why not clustering? Clustering is used to discover groups without a labeled target. Here, you explicitly have a target variable (minutes late) and want a direct prediction, so supervised regression is appropriate.
Segment customers into different groups to support a marketing department. ______
Customer segmentation for marketing is a classic clustering scenario. Typically, you have customer attributes and behaviors (purchase frequency, average order value, product categories, website activity, demographics) but you do not already have labels like “Segment A, Segment B.” Clustering algorithms group customers so that customers within a cluster are more similar to each other than to those in other clusters. Why not classification? Classification would require predefined segment labels in historical data (for example, you already know each customer’s segment and want to predict it for new customers). That’s not implied here. Why not regression? Regression predicts a numeric value (like lifetime value). Segmentation is about grouping, not predicting a continuous number.
Predict whether a student will complete a university course. ______
Predicting whether a student will complete a course is a yes/no outcome, which is a binary classification problem. The model uses features such as attendance, assignment submissions, grades, engagement metrics, and prior course history to predict one of two discrete classes: “complete” or “not complete.” Why not regression? Regression would be used if you were predicting a continuous value such as “final grade percentage” or “number of weeks until completion.” Even if you encoded yes/no as 1/0, the underlying task is still classification because the desired output is a category. Why not clustering? Clustering could group students by behavior patterns, but it would not directly answer the labeled question of completion without additional interpretation.
Want to practice all questions on the go?
Download Cloud Pass for free — includes practice tests, progress tracking & more.


Want to practice all questions on the go?
Get the free app
Download Cloud Pass for free — includes practice tests, progress tracking & more.
HOTSPOT - To complete the sentence, select the appropriate option in the answer area. Hot Area:
Data values that influence the prediction of a model are called ______.
Correct answer: B. features. Features are the input data values (columns) that the model uses to learn patterns and make predictions. For example, in a house-price model, square footage, number of bedrooms, and location are features because they influence the predicted price. Why the others are wrong: - A. dependant variables: In most ML terminology, the dependent variable is the target/output you are trying to predict (often synonymous with the label), not the inputs that influence the prediction. - C. identifiers: Identifiers (like CustomerID) uniquely identify records but typically should not be used as predictive inputs because they don’t represent generalizable patterns; they can also introduce data leakage. - D. labels: A label is the known outcome/target value used for training in supervised learning (for example, “price” or “spam/not spam”), not the input values that influence the prediction.
HOTSPOT - For each of the following statements, select Yes if the statement is true. Otherwise, select No. NOTE: Each correct selection is worth one point. Hot Area:
Forecasting housing prices based on historical data is an example of anomaly detection.
Forecasting housing prices based on historical data is not anomaly detection. The key phrase is “forecasting … based on historical data,” which indicates a time-series forecasting problem: you use past values (and possibly external variables like interest rates, inventory, seasonality) to predict future prices. The objective is to estimate expected future values, not to identify rare deviations. Anomaly detection would instead focus on finding unusual house prices or transactions that deviate from normal patterns (for example, a sale price far above comparable properties, potentially indicating data errors or fraud). In AI-900, forecasting and anomaly detection are distinct workload types: forecasting predicts what will happen; anomaly detection flags what is unusual compared to normal behavior. Therefore, the correct answer is No.
Identifying suspicious sign-ins by looking for deviations from usual patterns is an example of anomaly detection.
Identifying suspicious sign-ins by looking for deviations from usual patterns is a classic anomaly detection scenario. The goal is to detect rare or unusual events—such as logins from atypical locations, impossible travel (two distant sign-ins in a short time), unusual device fingerprints, abnormal login times, or spikes in failed attempts—relative to a baseline of normal user behavior. This aligns directly with the definition of anomaly detection: finding observations that differ significantly from the majority. It is often used in cybersecurity and fraud detection because anomalies can indicate compromised accounts or malicious activity. This is not primarily forecasting (predicting a future numeric value) or standard classification unless you already have labeled examples of “suspicious” vs. “normal.” Even when classification is used, the underlying concept in the prompt—“deviations from usual patterns”—maps to anomaly detection. Therefore, the correct answer is Yes.
Predicting whether a patient will develop diabetes based on the patient’s medical history is an example of anomaly detection.
Predicting whether a patient will develop diabetes based on the patient’s medical history is not anomaly detection; it is classification. The output is a discrete label (for example, “will develop diabetes: yes/no”), and the model is trained on historical patient records where the outcome is known. This is a supervised learning task. Anomaly detection would be more appropriate if the goal were to identify unusual patient measurements or rare patterns that deviate from a normal population (for example, detecting abnormal lab results that might indicate an undiagnosed condition). But the prompt is explicitly about predicting a specific outcome (diabetes development), which is a typical binary classification use case. In AI-900, remember: classification predicts categories, regression predicts numbers, forecasting predicts future time-based values, and anomaly detection flags rare deviations. Since the scenario is outcome prediction (yes/no), the correct answer is No.
DRAG DROP - Match the Microsoft guiding principles for responsible AI to the appropriate descriptions. To answer, drag the appropriate principle from the column on the left to its description on the right. Each principle may be used once, more than once, or not at all. NOTE: Each correct selection is worth one point. Select and Place:
Ensure that AI systems operate as they were originally designed, respond to unanticipated conditions, and resist harmful manipulation.
Correct: E (Reliability and safety). The description emphasizes that AI systems should behave consistently with their intended design, handle unexpected conditions, and resist harmful manipulation. These are classic reliability/safety concerns: robustness, fault tolerance, resilience to adversarial inputs, and safe operation under edge cases. Why others are wrong: - A (Accountability) is about who is responsible and ensuring oversight and governance, not the technical/operational robustness of the model. - B (Fairness) focuses on avoiding discriminatory outcomes across groups. - C (Inclusiveness) focuses on accessibility and ensuring solutions work for people with diverse abilities and needs. - D (Privacy and security) focuses on protecting data and controlling data usage, not primarily on model robustness and safe behavior.
Implementing processes to ensure that decisions made by AI systems can be overridden by humans.
Correct: A (Accountability). The key phrase is that AI decisions “can be overridden by humans.” This is human-in-the-loop governance: ensuring there is appropriate oversight, escalation paths, and the ability for people to intervene when an automated decision is incorrect or inappropriate. Accountability also implies clear ownership for outcomes and processes for auditing and remediation. Why others are wrong: - E (Reliability and safety) is about the system operating safely and robustly, but the explicit requirement for human override maps more directly to accountability and governance. - D (Privacy and security) concerns data protection and consent, not decision override. - B (Fairness) concerns bias and equitable outcomes. - C (Inclusiveness) concerns accessibility and designing for diverse users.
Provide consumers with information and controls over the collection, use, and storage of their data.
Correct: D (Privacy and security). The description is explicitly about giving consumers information and controls over the collection, use, and storage of their data. That aligns directly with privacy principles (transparency, consent, data minimization, appropriate retention) and security principles (protecting data at rest/in transit, access control). Why others are wrong: - A (Accountability) is about oversight and responsibility for AI outcomes, not user controls over personal data. - E (Reliability and safety) is about robust and safe system behavior. - B (Fairness) is about avoiding biased outcomes. - C (Inclusiveness) is about ensuring the solution is usable and beneficial for people with diverse needs, not data governance.
You need to develop a mobile app for employees to scan and store their expenses while travelling. Which type of computer vision should you use?
Semantic segmentation assigns a class label to each pixel (for example, separating road, sky, and cars in autonomous driving). It’s useful when you need precise object boundaries and pixel-level understanding. For expense receipts, pixel-level labeling is unnecessary; the goal is to extract text and key-value fields, which is better handled by OCR and document analysis services.
Image classification predicts a single label (or set of labels) for an entire image, such as “receipt,” “menu,” or “passport.” While classification could help route images (e.g., detect whether the photo is a receipt), it does not extract the merchant name, date, or total. You would still need OCR to read the receipt content and store expense details.
Object detection identifies and locates objects using bounding boxes (for example, detecting a laptop and a phone in a photo). It could potentially detect the presence and location of a receipt within a larger scene, enabling cropping. However, object detection does not convert printed text into machine-readable text, so it cannot capture expense amounts or vendor names without OCR.
Optical character recognition (OCR) is the correct choice because the app needs to read text from scanned expense receipts. OCR converts printed or handwritten text in an image into machine-readable text, which can then be stored in an expense system. This matches the scenario directly, since receipts primarily contain textual information such as merchant names, dates, and totals. In Azure AI-900 terms, whenever the requirement is to extract text from an image, OCR is the expected computer vision capability.
Core Concept: This question tests recognition of the correct computer vision workload for extracting text from images—Optical Character Recognition (OCR). In Azure AI, OCR is commonly delivered via Azure AI Vision (Image Analysis/Read) and is used to detect and extract printed or handwritten text from images and documents. Why the Answer is Correct: Employees scanning receipts while traveling need the app to capture the receipt image and then extract key expense details (merchant name, date, line items, totals, currency, taxes). The essential capability is converting the text on the receipt image into machine-readable text so it can be stored, searched, validated, and used for expense reporting workflows. OCR is specifically designed for this: it locates text regions and returns the recognized text with structure (lines/words) and confidence scores. Key Features / Best Practices: Azure’s OCR (often referred to as the “Read” capability in Azure AI Vision) supports printed and handwritten text, multiple languages, and returns bounding boxes and confidence values. For receipts, you typically combine OCR with post-processing: parsing totals/dates via rules or additional AI models, validating confidence thresholds, and handling image quality issues (blur, skew, poor lighting). From a solution perspective, store images in Azure Blob Storage and store extracted text/metadata in a database (e.g., Azure SQL or Cosmos DB). Consider privacy and compliance (PII on receipts) and secure access using managed identities and encryption. Common Misconceptions: Object detection can find a receipt in an image, but it doesn’t read the text. Image classification can label an image as “receipt” vs “not receipt,” but again doesn’t extract expense values. Face detection is unrelated unless you were verifying identity. Exam Tips: For AI-900, map the task to the workload: “read text from images/documents” => OCR (Azure AI Vision Read). “Find items and their locations” => object detection. “Assign a label to an image” => image classification. “Detect people/faces” => face detection. When the scenario mentions scanning forms, invoices, receipts, or extracting text, OCR is almost always the correct choice.
You need to create a training dataset and validation dataset from an existing dataset. Which module in the Azure Machine Learning designer should you use?
Select Columns in Dataset is used to include or exclude specific columns (features) from a dataset. It helps with feature selection, removing IDs, labels, or irrelevant fields, and shaping the schema before training. However, it does not partition rows into separate training and validation datasets, so it cannot create train/validation splits from one dataset.
Add Rows appends one dataset to another by stacking rows (a union/concatenation operation). It is useful when you have the same schema across datasets and want to increase the number of records (for example, combining monthly files). It does not split a dataset into training and validation subsets; it combines datasets instead.
Split Data is the correct module for creating training and validation (or test) datasets from an existing dataset. It partitions the rows into two outputs based on a specified ratio and can randomize the split for representative sampling. This is a standard step in Azure ML designer pipelines to train on one subset and evaluate on the other.
Join Data merges two datasets side-by-side based on a key column (similar to SQL joins). It is used to enrich data by bringing in additional features from another table. While important for data preparation, it does not create separate training and validation datasets from a single dataset; it combines datasets rather than splitting them.
Core concept: This question tests how to prepare data for machine learning by creating separate training and validation (or test) datasets from a single existing dataset. In Azure Machine Learning designer, this is a standard preprocessing step used to evaluate how well a model generalizes to unseen data. Why correct: The Split Data module is specifically designed to divide one dataset into two outputs. These outputs are commonly used as the training set and the validation or test set. It supports percentage-based splitting and can randomize the rows so each subset is more representative of the original dataset. Key features: Split Data lets you control how much of the data goes to each output, such as 70/30 or 80/20. It can also use randomized splitting and support reproducible results through consistent configuration. In a typical Azure ML designer workflow, one output is used to train the model and the other is used to score and evaluate it. Common misconceptions: Some learners confuse selecting columns with splitting data, but Select Columns in Dataset only changes which fields are included and does not create separate subsets of rows. Add Rows and Join Data are both used to combine datasets, not divide one dataset into training and validation portions. Exam tips: For AI-900, remember that when a question asks about creating training and validation or test datasets from one dataset, the correct module is Split Data. If the question is about choosing features, think Select Columns in Dataset. If it is about combining datasets, think Add Rows or Join Data.
You are designing an AI system that empowers everyone, including people who have hearing, visual, and other impairments. This is an example of which Microsoft guiding principle for responsible AI?
Fairness is about ensuring AI systems treat people equitably and do not create or reinforce bias. It focuses on outcomes across groups (for example, similar error rates or decision impacts across demographics). While fairness can include considering underserved groups, the scenario is explicitly about accessibility for people with impairments, which is more directly the inclusiveness principle.
Inclusiveness is the principle that AI should empower and engage everyone, including people with disabilities. It emphasizes accessibility and designing AI experiences that accommodate diverse abilities (e.g., captions for hearing impairments, screen-reader-friendly outputs for visual impairments, alternative interaction modes). The scenario’s wording matches inclusiveness almost verbatim.
Reliability and safety focuses on whether an AI system performs consistently under expected and unexpected conditions, and whether it avoids causing harm through failures. Examples include robustness to edge cases, monitoring, fallback behaviors, and safe deployment practices. The scenario is not about system stability or safety controls; it is about accessibility and broad usability.
Accountability is about having appropriate governance, oversight, and responsibility for AI systems. It includes human supervision, auditability, and clear ownership for decisions and impacts. While accountability is important for any AI solution, the scenario’s emphasis is on enabling people with impairments to use the system, which aligns with inclusiveness rather than governance.
Core Concept: This question tests Microsoft’s Responsible AI guiding principles. These principles describe how AI systems should be designed and operated to benefit people and society. In AI-900, you’re expected to recognize the intent of each principle from short scenarios. Why the Answer is Correct: Designing an AI system that “empowers everyone, including people who have hearing, visual, and other impairments” directly maps to the principle of inclusiveness. Inclusiveness is about ensuring AI systems are accessible and usable by people with diverse abilities, backgrounds, and needs, including people with disabilities. It emphasizes designing experiences that remove barriers (for example, supporting screen readers, captions, alternative input methods, and assistive technologies) so that the benefits of AI are broadly shared. Key Features / Best Practices: Inclusiveness in AI often involves: - Accessibility-by-design: ensuring interfaces and outputs work for users with visual/hearing/mobility impairments (captions, transcripts, high-contrast UI, keyboard navigation). - Multimodal experiences: offering text, audio, and visual alternatives (e.g., speech-to-text and text-to-speech). - Inclusive data and testing: validating the system with diverse user groups and assistive technology scenarios. This aligns with Azure Well-Architected Framework principles such as Reliability and Operational Excellence when you incorporate inclusive testing and monitoring, but the guiding principle being tested is specifically Responsible AI “Inclusiveness.” Common Misconceptions: - Fairness can sound similar because it’s also about avoiding harm to groups, but fairness focuses on equitable outcomes and avoiding bias/discrimination in model decisions. - Reliability and safety focuses on robustness, failure modes, and safe operation, not accessibility. - Accountability is about governance, human oversight, and responsibility for outcomes. Exam Tips: For AI-900, look for keywords: - “Accessible,” “people with disabilities,” “empower everyone,” “assistive” → Inclusiveness. - “Bias,” “equal treatment,” “disparate impact” → Fairness. - “Robust,” “safe,” “resilient,” “fails gracefully” → Reliability and safety. - “Human in the loop,” “audit,” “governance,” “responsible party” → Accountability.
DRAG DROP - Match the machine learning tasks to the appropriate scenarios. To answer, drag the appropriate task from the column on the left to its scenario on the right. Each task may be used once, more than once, or not at all. NOTE: Each correct selection is worth one point. Select and Place:
______ Examining the values of a confusion matrix
Examining the values of a confusion matrix is part of model evaluation. A confusion matrix summarizes classification outcomes (true positives, false positives, true negatives, false negatives) on validation/test data, helping you assess performance and calculate metrics like precision and recall. Why not the others: - Feature engineering (A) changes/creates inputs; it doesn’t assess predictions. - Feature selection (B) chooses which inputs to keep. - Model training (E) is the learning/fitting step; the confusion matrix is typically produced after training when scoring a dataset. - Model deployment (C) is packaging and hosting the model for real use (endpoint, container, etc.), not analyzing metrics.
______ Splitting a date into month, day, and year fields
Splitting a date into month, day, and year fields is feature engineering because you are transforming a raw attribute into multiple derived features that may better capture patterns. Many models can’t directly learn useful relationships from a single date string/timestamp, but can learn seasonality (month), day-of-month effects, or year trends when those components are explicitly provided. Why not the others: - Feature selection (B) would be choosing whether to use the date (or derived fields) at all, not performing the split. - Model training (E) is fitting the algorithm after features are prepared. - Model evaluation (D) is measuring performance. - Model deployment (C) is operationalizing the trained model.
______ Picking temperature and pressure to train a weather model
Picking temperature and pressure to train a weather model is feature selection. You are deciding which candidate variables (features) to include as inputs to the model. Good feature selection can reduce overfitting, improve generalization, and lower training cost by removing irrelevant or redundant inputs. Why not the others: - Feature engineering (A) would be transforming temperature/pressure (for example, normalizing values, creating rolling averages, or computing pressure change over time). - Model training (E) is the step where the algorithm learns from the selected features. - Model evaluation (D) happens after training to validate performance. - Model deployment (C) is publishing the model for inference (for example, an Azure ML endpoint).
You are developing a natural language processing solution in Azure. The solution will analyze customer reviews and determine how positive or negative each review is. This is an example of which type of natural language processing workload?
Language detection identifies the language of input text (for example, English vs. French) and often returns a language code with a confidence score. It is useful as a preprocessing step when you have multilingual reviews, but it does not determine whether the review is positive or negative. Therefore, it does not match the requirement to measure sentiment.
Sentiment analysis is the NLP task that evaluates the emotional tone of text and determines whether it is positive, negative, neutral, or mixed. For customer reviews, it can provide an overall sentiment label and confidence scores, and sometimes sentence-level sentiment to pinpoint which parts of the review drive the rating. This directly matches the scenario described.
Key phrase extraction pulls out the most important words or phrases from text (for example, “battery life,” “customer service,” “delivery time”). It helps summarize what topics customers mention, but it does not assess whether those topics are discussed positively or negatively. It’s often paired with sentiment analysis, but it is not the same workload.
Entity recognition (named entity recognition) identifies and categorizes entities such as people, organizations, locations, dates, or product names. It’s useful for understanding what is being discussed (e.g., “Contoso,” “Seattle”), but it does not measure opinion polarity. As a result, it does not satisfy the requirement to determine positivity or negativity.
Core Concept: This question tests recognition of common Natural Language Processing (NLP) workloads in Azure, specifically the ability to determine whether text expresses a positive, negative, neutral, or mixed opinion. In Azure AI services, this capability is provided by Sentiment Analysis (part of Azure AI Language, formerly Text Analytics). Why the Answer is Correct: The solution analyzes customer reviews and determines “how positive or negative each review is.” That is the definition of sentiment analysis: classifying and scoring the emotional tone/opinion in text. In Azure AI Language, sentiment analysis typically returns an overall document sentiment label (positive/neutral/negative/mixed) and confidence scores, and can also provide sentence-level sentiment for more granular insights. Key Features / How It’s Used: Sentiment analysis is commonly applied to customer feedback, social media monitoring, and support tickets to measure satisfaction and detect issues early. In Azure AI Language, you submit text (single or batch), choose the language (or combine with language detection), and receive sentiment labels and scores. Best practices include cleaning noisy text, handling sarcasm/irony limitations, and evaluating results on your domain-specific data. From an Azure Well-Architected Framework perspective, consider reliability (retry policies, regional deployment), security (avoid sending sensitive data; use private endpoints where available), and cost optimization (batch requests, right-size throughput). Common Misconceptions: Language detection can be involved in an end-to-end pipeline, but it only identifies the language (e.g., English, Spanish) and does not measure positivity/negativity. Key phrase extraction finds important terms, not sentiment. Entity recognition identifies names of people/places/organizations, not opinion. Exam Tips: For AI-900, map the wording to the workload: - “positive/negative opinion” => sentiment analysis - “what language is this?” => language detection - “main talking points” => key phrase extraction - “people/places/brands” => entity recognition Also remember that these are prebuilt NLP capabilities in Azure AI Language rather than custom model training in most fundamentals-level questions.
You need to provide content for a business chatbot that will help answer simple user queries. What are three ways to create question and answer text by using QnA Maker? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.
Correct. QnA Maker can generate Q&A pairs by ingesting an existing webpage (commonly an FAQ page). You provide a URL, and the service extracts question/answer content or structured sections to form knowledge base entries. This is a primary, exam-relevant method for quickly bootstrapping a chatbot knowledge base from existing documentation.
Incorrect. Automated machine learning (AutoML) is used to train ML models (classification, regression, forecasting) in Azure Machine Learning, not to create a QnA Maker knowledge base. QnA Maker focuses on authoring Q&A pairs and using built-in ranking/matching; it does not require you to run AutoML on a file of questions to create the Q&A content.
Correct. You can manually enter questions and answers in the QnA Maker portal (or via APIs). Manual authoring is important for curating high-quality responses, adding alternate questions (paraphrases), and refining answers. This is a straightforward way to create or supplement a knowledge base when you don’t have a suitable source document or webpage.
Incorrect. Connecting a bot to a channel (such as Cortana, Teams, or Web Chat) is a deployment/integration step handled through Azure Bot Service and Bot Framework channels. It does not create Q&A pairs in QnA Maker. Asking questions via a channel is how users interact with the bot after publishing, not how the knowledge base content is authored.
Correct. QnA Maker supports importing chit-chat content from predefined datasets. This adds common conversational Q&A (greetings, pleasantries, small talk) to improve the user experience beyond business FAQs. It is explicitly a content-creation method within QnA Maker authoring and is commonly tested in AI-900.
Core Concept: This question tests knowledge of how to populate a knowledge base in QnA Maker (part of Azure Cognitive Services for building FAQ-style question answering for chatbots). In AI-900 terms, this sits in Natural Language Processing (NLP): creating question-and-answer pairs that a bot can use to respond to user queries. Why the Answer is Correct: QnA Maker supports multiple authoring methods to create Q&A pairs. You can (1) extract Q&A content from existing sources like webpages, (2) manually author Q&A pairs directly in the portal, and (3) add “chit-chat” datasets (prebuilt conversational Q&A) to make the bot respond naturally to greetings and small talk. These correspond to A, C, and E. Key Features / How it Works: - Source ingestion: QnA Maker can ingest content from URLs (FAQ pages), documents (such as PDFs/Word), and other supported sources, then extract question/answer pairs or structured content. - Manual authoring: You can add, edit, and curate Q&A pairs, including alternate questions (synonyms) to improve matching. - Chit-chat: You can import a predefined chit-chat personality dataset (for example, “friendly” or “professional”) to handle common conversational phrases. Operationally, after authoring you must train and publish the knowledge base so it can be queried by a bot or application endpoint. Common Misconceptions: Option B is tempting because “training a model” sounds like ML, but QnA Maker is not created via Automated ML on a dataset of questions; it’s a managed Q&A matching service with authoring and ranking, not an AutoML workflow. Option D confuses deployment channels with content creation: connecting to Cortana (or any channel) is about where the bot is available, not how Q&A text is authored. Exam Tips: For AI-900, remember QnA Maker authoring methods: URL/document ingestion, manual entry, and built-in chit-chat. Also distinguish “creating content” (knowledge base authoring) from “publishing/integrating” (Bot Framework channels, endpoints). If an option describes channels or AutoML training, it’s usually not a QnA Maker authoring method.