
50問と45分の制限時間で実際の試験をシミュレーションしましょう。AI検証済み解答と詳細な解説で学習できます。
AI搭載
すべての解答は3つの主要AIモデルで交差検証され、最高の精度を保証します。選択肢ごとの詳細な解説と深い問題分析を提供します。
Your company is designing a data store that will contain student data. The data has the following format. StudentNumber | StudentInformation 7634634 First name: Ben Last: Smith Preferred Name: Benjamin 7634634 First Name: Dominik Last Name: Paiha Email Address: dpaiha@contoso.com MCP ID: 931817 7634636 First Name: Reshma Last Name: Patel Phone number: 514-555-1101 7634637 First Name: Yun-Feng Last Name: Peng Which type of data store should you use?
Graph databases are designed for highly connected data and relationship traversal (e.g., students enrolled in courses, friendships, prerequisites). The prompt shows no relationships—only a student identifier with varying attributes. Using a graph model would add complexity without benefit unless the workload requires queries like “find students connected through shared classes” or multi-hop relationship exploration.
Key/value fits because StudentNumber is a natural unique key and StudentInformation can be stored as a flexible value (commonly JSON). Each student can have different fields without schema changes, and the primary access pattern is efficient point lookup by key. This aligns with services like Azure Cosmos DB (NoSQL) or Azure Table Storage for simple, scalable, low-latency retrieval by identifier.
Object storage (e.g., Azure Blob Storage) is optimized for storing and serving unstructured files such as documents, images, backups, and logs. While you could store each student profile as a separate blob, object stores are not ideal for frequent record-level updates and key-based queries with low-latency transactional access patterns compared to a key/value or document database.
Columnar stores (e.g., data warehouses/columnstore formats like Parquet) are optimized for analytical workloads that scan large datasets and aggregate across columns (reporting, BI). The student data shown is operational/profile-style data with variable attributes and likely point reads by StudentNumber. Columnar storage would be inefficient and overly complex for this transactional, semi-structured use case.
Core concept: This question tests selecting an appropriate non-relational (NoSQL) data model based on the shape of the data. The sample shows a StudentNumber paired with a set of attributes that vary per student (some have Email and MCP ID, others have Phone, others only names). This is semi-structured data with a natural unique identifier. Why the answer is correct: A key/value data store is the best fit because each record can be stored and retrieved by a single key (StudentNumber), and the value can be an opaque blob (often JSON) containing the student’s attributes. The schema can vary per student without requiring table alterations or fixed columns. This aligns with common NoSQL patterns: fast point reads/writes by key, flexible schema, and simple access patterns. Key features / Azure context: In Azure, this pattern maps well to Azure Cosmos DB (NoSQL API) or Azure Table Storage, where the StudentNumber can be the partition key and/or row key (or the Cosmos DB item id plus a partition key). Best practices include choosing a partition key that supports even distribution and your query patterns (e.g., StudentNumber if access is primarily by student). From an Azure Well-Architected Framework perspective, key/value supports performance efficiency (low-latency point reads), operational excellence (no schema migrations for new optional fields), and reliability (managed, replicated services). Common misconceptions: Graph may seem plausible because “student information” could relate to courses, teachers, etc., but the provided data shows no relationships—just per-student attributes. Object storage is for files/blobs (documents, images) rather than queryable records keyed for frequent lookups. Columnar storage is optimized for analytics scans and aggregations across many rows/columns, not transactional lookups of individual student profiles. Exam tips: When you see “ID -> varying attributes” and the main operation is retrieve/update by that ID, think key/value (or document). If the question offered “document,” that would also be a strong candidate; with the given options, key/value is the closest match. Choose graph only when relationships and traversals are central, and choose columnar for large-scale analytical workloads (BI, reporting) rather than operational profile storage.
DRAG DROP - Match the security components to the appropriate scenarios. To answer, drag the appropriate component from the column on the left to its scenario on the right. Each component may be used once, more than once, or not at all. NOTE: Each correct match is worth one point. Select and Place:
______ Prevent access to an Azure SQL database from another network.
Correct answer: B. Firewall. To prevent access to an Azure SQL database from another network, you use network access controls such as the Azure SQL server-level firewall rules (allow/deny public IP ranges) and/or virtual network rules with Private Link/Private Endpoint to restrict connectivity to a specific VNet. These controls operate at the network perimeter and determine whether traffic from a given source network can reach the database endpoint. Why not A (Authentication): Authentication controls who can log in after a connection is established. It does not stop a network from reaching the service endpoint; it only rejects invalid identities. Why not C (Encryption): Encryption protects confidentiality of data at rest or in transit, but it does not block network connectivity. An encrypted database can still be reachable from other networks unless firewall/VNet restrictions are configured.
______ Support Azure Active Directory (Azure AD) sign-ins to an Azure SQL database.
Correct answer: A. Authentication. Supporting Azure Active Directory (Azure AD) sign-ins to Azure SQL Database is an authentication capability. Azure AD authentication enables users and applications to authenticate using Azure AD identities (users, groups, managed identities, service principals) instead of (or in addition to) SQL logins. This is commonly paired with modern security features like MFA and Conditional Access, improving identity security and aligning with best practices. Why not B (Firewall): A firewall can restrict where connections come from, but it does not provide Azure AD sign-in capability. You could allow traffic through the firewall and still not support Azure AD authentication if it’s not configured. Why not C (Encryption): Encryption secures data, but it does not determine how users authenticate. Azure AD sign-in is about identity verification and token-based authentication, not data confidentiality.
______ Ensure that sensitive data never appears as plain text in an Azure SQL database.
Correct answer: C. Encryption. Ensuring sensitive data never appears as plain text in an Azure SQL database points to encryption. In Azure SQL, this is most strongly associated with technologies like Always Encrypted, which keeps sensitive column data encrypted end-to-end so that the SQL engine never sees plaintext values (encryption/decryption occurs in the client driver). More generally, encryption mechanisms (TDE, TLS) protect data at rest and in transit, but the “never appears as plain text” wording aligns best with Always Encrypted-style protection. Why not A (Authentication): Authentication controls access by identity, but authenticated users (or admins) could still view plaintext data if it is stored unencrypted. Why not B (Firewall): Firewalls restrict network access but do not prevent plaintext storage or exposure once access is granted.
Which storage solution supports role-based access control (RBAC) at the file and folder level?
Azure Disk Storage provides block-level storage for Azure VMs (managed disks). It does not expose a filesystem service where Azure can apply RBAC at the file and folder level. Access is controlled through VM access and disk-level permissions (management plane) rather than data-plane file/folder authorization. You can format a disk with NTFS/ext4 inside a VM, but that’s OS-level permissions, not Azure Storage RBAC on files/folders.
Azure Data Lake Storage Gen2 is the Azure storage service that supports real directories and files through its hierarchical namespace. That structure allows permissions to be applied at the directory and file level, which is not possible in the same way with flat object storage. In practice, Azure RBAC is used for broader authorization and POSIX-style ACLs are used for fine-grained control on folders and files. This makes ADLS Gen2 the best and expected answer for storage scenarios requiring granular filesystem-style security.
Azure Blob storage is primarily an object store with a flat namespace, even though blob names can include slash-delimited prefixes that look like folders. Those prefixes are virtual and do not provide true directory objects with independent file/folder permissions. Azure RBAC can grant access to blob data, but not true folder-level authorization in the same filesystem sense tested here. The granular file and directory permission model is associated with ADLS Gen2, not standard Blob storage.
Azure Queue storage is a messaging service for storing and retrieving messages. It does not have a concept of files and folders, so file/folder-level RBAC is not applicable. Authorization is typically via storage account keys, SAS tokens, or Azure AD with appropriate roles for queue operations, but it’s still at the queue/message operation level rather than filesystem-style permissions.
Core concept: This question tests which Azure storage service supports granular authorization for directories and files. In Azure, true file- and folder-level permissions require a hierarchical filesystem model rather than flat object storage or block storage. Why the answer is correct: Azure Data Lake Storage Gen2 supports a hierarchical namespace, which introduces real directories and files. This enables POSIX-style access control lists (ACLs) that can be applied directly to folders and files. Azure RBAC can be used to grant broader data access roles, but the file- and folder-level granularity itself is provided by ACLs in ADLS Gen2. Key features / configurations / best practices: - Use Azure Data Lake Storage Gen2 with hierarchical namespace enabled for directory-aware storage. - Use Microsoft Entra ID identities together with Azure RBAC for coarse-grained access and ACLs for fine-grained file/folder permissions. - Apply least-privilege principles by assigning only the required storage roles and then narrowing access with ACLs. Common misconceptions: - Azure Blob storage supports Azure RBAC for blob data access, but it does not provide true folder-level security in the same way unless using ADLS Gen2 hierarchical namespace features. - Azure Disk Storage is attached block storage for virtual machines, not a managed filesystem service with Azure-enforced file/folder permissions. - Queue storage is for messages and has no file or folder structure. Exam tips: When a question mentions file- or folder-level access control in Azure storage, think of Azure Data Lake Storage Gen2 and its hierarchical namespace with ACLs. If the wording says RBAC, remember that exams sometimes simplify the distinction, but the granular enforcement in ADLS Gen2 is done through ACLs.
Which type of non-relational data store supports a flexible schema, stores data as JSON files, and stores the all the data for an entity in the same document?
Document stores keep each entity as a single document (commonly JSON/BSON) and allow different documents to have different fields (schema flexibility). This matches the prompt exactly: flexible schema, JSON representation, and storing all data for an entity together. Azure Cosmos DB (for NoSQL) is the common Azure example used in DP-900 for document databases.
Columnar (column-family/columnar) storage organizes data by columns rather than rows/documents and is typically optimized for analytics (large scans, aggregations, compression). While some column-family systems can handle sparse attributes, they do not primarily store each entity as a single JSON document. In Azure exam context, columnar is more associated with analytical engines and data warehousing patterns.
Graph databases store data as nodes (vertices) and relationships (edges), optimized for relationship traversal (e.g., social networks, recommendations). Graph models can be schema-flexible, but the defining feature is relationship-centric querying (traversals), not storing an entire entity as one JSON document. In Azure, Cosmos DB also supports a graph API, but that’s distinct from document modeling.
Time series databases are optimized for high-ingest, timestamped data (metrics, IoT telemetry) and time-window queries (downsampling, rollups). They may store records in various formats, but the defining characteristic is time-based indexing and querying, not general-purpose JSON documents containing all attributes of an entity. Azure examples include Azure Data Explorer and time-series patterns in Cosmos DB.
Core concept: This question tests recognition of non-relational (NoSQL) data models, specifically the document data model. In Azure fundamentals (DP-900), you’re expected to map workload characteristics (flexible schema, JSON, entity stored together) to the correct NoSQL store type. Why the answer is correct: A document data store is designed to store semi-structured data as “documents,” commonly represented as JSON (or BSON). A key property is that all attributes for a single entity are stored together in one document, rather than being split across multiple tables/rows as in relational systems. Document stores support a flexible schema: different documents in the same collection can have different fields, which is ideal when data evolves frequently or varies by entity. Key features and best practices: Document databases typically organize data into collections/containers and documents. They enable fast retrieval of an entire entity because the entity is co-located in one document, reducing the need for joins. In Azure, Azure Cosmos DB (for NoSQL) is the flagship document-oriented service, storing JSON documents and supporting schema flexibility. A best practice aligned with the Azure Well-Architected Framework (Performance Efficiency and Reliability) is to model data around access patterns: embed related data that is read together, and choose an appropriate partition key to scale horizontally and avoid hot partitions. Common misconceptions: Columnar stores can feel “flexible” because they handle sparse columns, but they are optimized for analytics and store data by columns, not as JSON documents per entity. Graph stores also have flexible schemas, but they model relationships as vertices/edges rather than a single JSON document per entity. Time series stores handle timestamped measurements and are optimized for time-based queries, not general entity documents. Exam tips: When you see “JSON,” “flexible schema,” and “all data for an entity in the same document,” think document store (e.g., Cosmos DB for NoSQL). If the question emphasizes relationships and traversals, think graph. If it emphasizes analytical scans and compression by column, think columnar. If it emphasizes telemetry over time and windowing, think time series.
HOTSPOT - For each of the following statements, select Yes if the statement is true. Otherwise, select No. NOTE: Each correct selection is worth one point. Hot Area:
Normalization involves eliminating relationships between database tables.
Normalization does not eliminate relationships between tables; it typically increases the need for well-defined relationships. When you normalize, you often split a wide table into multiple tables so that each table represents a single subject (entity) and each non-key attribute depends on the key. After splitting, you connect the tables using primary keys and foreign keys (relationships) to reconstruct the full view of the data through joins. For example, moving customer address details into a separate Address table and referencing it from a Customer table creates (or reinforces) a relationship rather than removing one. Eliminating relationships would be contrary to relational design and would usually lead to duplicated data or loss of referential integrity. Therefore, the statement is false, so the correct choice is No.
Normalizing a database reduces data redundancy.
A primary goal of normalization is to reduce data redundancy (duplicate storage of the same facts). By organizing data into separate tables based on dependencies, you avoid repeating the same values across many rows. This reduces update anomalies—for instance, if a customer’s phone number is stored in multiple places, changing it in one place but not another creates inconsistency. In normalized designs, shared facts are stored once and referenced via keys. For example, storing product details in a Product table and referencing ProductID from an OrderLine table prevents repeating product name and description on every order line. While normalization can increase the number of tables and joins, it generally decreases redundant data storage and improves maintainability. Therefore, the statement is true, so the correct choice is Yes.
Normalization improves data integrity.
Normalization improves data integrity by reducing the likelihood of inconsistent or invalid data. When the same fact is stored in multiple locations (redundancy), it’s easy for those copies to diverge over time, harming consistency. Normalization centralizes each fact in one place and uses keys and constraints to enforce correct relationships. For example, with a properly normalized schema, a foreign key constraint can ensure that every Order references a valid Customer. This supports referential integrity. Additionally, by aligning columns with the correct entity and dependency rules (e.g., 2NF/3NF principles), you reduce anomalies where inserts, updates, or deletes accidentally create orphaned or contradictory data. Although integrity also depends on implementing constraints (PK, FK, UNIQUE, CHECK) and correct transactions, normalization is a foundational design practice that makes it easier to enforce and maintain integrity. Therefore, the statement is true, so the correct choice is Yes.
外出先でもすべての問題を解きたいですか?
Cloud Passを無料でダウンロード — 模擬試験、学習進捗の追跡などを提供します。
HOTSPOT - For each of the following statements, select Yes if the statement is true. Otherwise, select No. NOTE: Each correct selection is worth one point. Hot Area:
Processing salary payments once a month is an example of a batch workload.
Yes. Processing salary payments once a month is a classic batch workload. The defining characteristics are (1) a fixed schedule (monthly), (2) a bounded set of records to process (all employees/payroll changes for the pay period), and (3) no requirement to process each individual payroll event the moment it occurs. Batch systems prioritize reliability, auditability, and throughput over low latency. In practice, payroll is often run as a scheduled job that aggregates time/attendance, applies deductions and taxes, and produces outputs (bank files, payslips, ledger entries) in one controlled run. Why not streaming? Streaming would imply continuous processing of payroll-related events in near real time (for example, calculating pay instantly as hours are logged). That is not the typical business requirement for monthly salary payments, where results are needed at the end of the pay cycle.
A wind turbine that sends 50 sensor readings per second is an example of a streaming workload.
Yes. A wind turbine sending 50 sensor readings per second is an example of a streaming workload because it produces a continuous, high-velocity flow of events. At 50 readings/second, the data arrives frequently enough that it is naturally modeled as an event stream, and it commonly supports near-real-time use cases such as anomaly detection, predictive maintenance, operational dashboards, and alerting. In Azure terms, this pattern aligns with IoT telemetry ingestion (Azure IoT Hub or Event Hubs), real-time processing (Azure Stream Analytics, Databricks/Synapse Spark streaming), and storage/analytics for hot and warm paths (Azure Data Explorer, Cosmos DB, or a time-series optimized store). It’s not batch because you typically wouldn’t wait hours or days to process turbine telemetry if the goal is to detect faults quickly; the value comes from low-latency processing.
A home electricity meter that sends readings once a day to an energy provider is an example of a streaming workload.
No. A home electricity meter that sends readings once a day is better described as a batch (or periodic) workload rather than streaming. Although the data may be transmitted electronically, the key factor is the low frequency and the implied lack of near-real-time processing requirements. Daily readings are typically used for billing, consumption summaries, and trend reporting—use cases that tolerate higher latency and are commonly processed in scheduled jobs. Why not streaming? Streaming implies continuous ingestion and processing of events as they occur, usually with seconds-to-minutes latency. Once-per-day telemetry is not a continuous high-velocity stream; it’s a periodic snapshot. In Azure, this would often be ingested and then processed via scheduled ETL/ELT (for example, Data Factory) rather than real-time stream processing services, unless there are additional requirements (like immediate outage detection) that would necessitate more frequent reporting.
HOTSPOT - To complete the sentence, select the appropriate option in the answer area. Hot Area:
In Azure Data Factory, you can use ______ to orchestrate pipeline activities that depend on the output of other pipeline activities.
Correct answer: A. a control flow In Azure Data Factory, control flow is the capability within a pipeline that orchestrates activity execution order and dependencies. You use dependency conditions (for example, an activity runs only when a prior activity Succeeds) and control activities such as If Condition, Switch, ForEach, Until, Wait, and Execute Pipeline to coordinate steps that rely on earlier outputs or completion states. This is exactly what “orchestrate pipeline activities that depend on the output of other pipeline activities” describes. Why the others are wrong: - B. Dataset: defines the data structure and location (tables, files, folders) used by activities; it does not control execution order. - C. Linked service: stores connection information (server, authentication, endpoints) to a data store or compute; not orchestration. - D. Integration runtime: provides the compute infrastructure for data movement and transformation (Azure IR, Self-hosted IR); it executes work but does not define dependency logic between activities.
HOTSPOT - You have the following JSON document.
"customer" : {
"first name" : "Ben",
"last name" : "Smith",
"address" : {
"line 1" : "161 Azure Ln",
"line 2" : "Palo Alto",
"ZIP code" : "54762"
},
"social media": [
{
"service" : "twitter",
"handle" : "@bensmith"
},
{
"service" : "linkedin",
"handle" : "bensmith"
}
],
"phone numbers": [
{
"type" : "mobile",
"number" : "555-555-555"
}
]
}
Use the drop-down menus to select the answer choice that completes each statement based on the information presented in the JSON document. NOTE: Each correct selection is worth one point. Hot Area:
Customer is ______.
Customer is a root object in the context of the provided JSON snippet. The snippet shows a top-level property named "customer" whose value is an object (it uses curly braces { }). Because it is not shown as being contained inside another named property (for example, inside a larger object like { "order": { "customer": { ... } } }), it is treated as the top-level/root object for this document. Why not A (nested array): "customer" is not in square brackets [ ], so it is not an array. Why not B (nested object): Although "customer" contains nested structures, the question’s intent is to classify "customer" itself. In this snippet, it is the primary/top-level object being described, so it is best classified as the root object.
Address is ______.
Address is a nested object. Inside the "customer" object, the "address" property’s value is enclosed in curly braces { }, which indicates an object. Because it exists within the "customer" object rather than at the top level of the document, it is nested. Why not A (nested array): "address" is not defined with square brackets [ ], and it does not contain multiple elements; it is a single object with properties like "line 1", "line 2", and "ZIP code". Why not C (root object): A root object is the top-level object of the JSON document. "address" is clearly a child property under "customer", so it cannot be the root.
Social media is ______.
Social media is a nested array. The "social media" property is defined using square brackets [ ], which indicates an array. Each element in that array is an object (each has curly braces { } with "service" and "handle"). Because the array is contained within the "customer" object, it is nested. Why not B (nested object): While each item inside the array is an object, the "social media" property itself is an array of objects, not a single object. Why not C (root object): "social media" is not at the top level; it is a property inside "customer", so it is not a root object.
You need to gather real-time telemetry data from a mobile application. Which type of workload describes this scenario?
OLTP is designed for high volumes of short, atomic transactions (insert/update/select) with strong consistency and ACID guarantees, typically in relational databases. A mobile app can generate OLTP data (e.g., user profile updates or purchases), but “real-time telemetry” is event streaming rather than transactional processing. OLTP focuses on operational systems, not continuous event ingestion and windowed analytics.
Batch workloads process data in periodic, scheduled runs over a bounded dataset (for example, nightly ETL, daily aggregations, or monthly billing). Telemetry can be stored and later processed in batch, but the question explicitly says “real-time,” which implies low-latency processing as events arrive. Batch is the opposite of continuous, near-immediate analytics.
MPP (massively parallel processing) refers to a distributed compute architecture used for large-scale analytical queries by splitting work across many nodes (common in data warehouses). It’s not primarily a workload classification for real-time telemetry ingestion. MPP can analyze large telemetry datasets after ingestion, but the scenario is about continuous real-time data flow, which is better described as streaming.
Streaming workloads handle continuous, unbounded flows of events and process them with low latency. Real-time telemetry from a mobile application (app events, performance metrics, location updates) is a classic streaming scenario. In Azure, this commonly maps to Event Hubs/IoT Hub for ingestion and Stream Analytics/Spark for real-time processing, enabling dashboards, alerts, and near-real-time insights.
Core concept: This question tests your ability to classify data workloads, specifically recognizing a streaming analytics workload. Real-time telemetry from a mobile app is a continuous flow of events (clicks, location pings, performance metrics, custom events) that must be ingested and processed with low latency. Why the answer is correct: “Real-time telemetry” implies data arrives continuously and you want near-immediate processing/insights (seconds or sub-seconds). That is the definition of a streaming workload: unbounded data, event-by-event or micro-batch processing, and time-sensitive analytics/alerting. In Azure, typical services for this pattern include Azure Event Hubs or IoT Hub for ingestion, Azure Stream Analytics or Spark Structured Streaming for processing, and sinks like Azure Data Explorer, Cosmos DB, or a data lake for storage and analysis. Key features and best practices: Streaming workloads emphasize high-throughput ingestion, partitioning, and scalability (e.g., Event Hubs partitions, consumer groups). They often use windowing (tumbling/hopping/sliding windows) to aggregate over time, handle late/out-of-order events, and support real-time dashboards and alerts. From an Azure Well-Architected Framework perspective, you design for reliability (durable ingestion, retry policies), performance efficiency (right-size throughput units/partitions), and cost optimization (avoid over-provisioning, use autoscale where available). Common misconceptions: OLTP can involve “online” apps, but OLTP refers to transactional reads/writes to a relational database with ACID properties (orders, payments, inventory), not continuous event streams. Batch workloads also process telemetry, but batch implies periodic processing of bounded datasets (hourly/daily jobs), not real-time. MPP describes a compute architecture for large-scale parallel analytics (e.g., dedicated SQL pools) rather than the workload type of real-time ingestion. Exam tips: When you see keywords like “real-time,” “telemetry,” “events,” “sensor data,” “clickstream,” or “continuous ingestion,” think streaming. When you see “daily/weekly processing,” “ETL at night,” or “historical reports,” think batch. When you see “transactions,” “orders,” “updates,” and “ACID,” think OLTP.
HOTSPOT - To complete the sentence, select the appropriate option in the answer area. Hot Area:
The massively parallel processing (MPP) engine of Azure Synapse Analytics ______
Correct answer: A. distributes processing across compute nodes. In Azure Synapse Analytics dedicated SQL pool, the MPP engine is designed to execute queries by splitting the work into parallel tasks that run across multiple compute nodes. The control node receives the client query, performs parsing and optimization, and then coordinates execution by distributing the query plan steps to the compute nodes. Each compute node processes its assigned data distributions in parallel, which is the core benefit of MPP for large analytics workloads. Why the others are wrong: - B (distributes processing across control nodes): Synapse does not distribute processing across multiple control nodes; the control node primarily orchestrates. - C (redirects client connections across compute nodes): Clients connect to the control node endpoint; compute nodes are not used for client connection redirection. - D (redirects client connections across control nodes): There isn’t a pool of control nodes for connection redirection in this model; the control node is the coordination point.
外出先でもすべての問題を解きたいですか?
無料アプリを入手
Cloud Passを無料でダウンロード — 模擬試験、学習進捗の追跡などを提供します。