1.
Which statement is correct about Horizontal Partitioning?
2.
...................... supports up to 100 TB of data and provides high throughput and performance, as well as rapid scaling to adapt to the workload requirements.
3.
You need to design an Azure Synapse Analytics dedicated SQL pool that can return an employee record from a given point in time, maintains the latest employee information, and minimizes query complexity. How should you model the employee data?
4.
You are designing an Azure Stream Analytics job to process incoming events from sensors in retail environments. You need to process the events to produce a running average of shopper counts during the previous 15 minutes, calculated at five-minute intervals. Which type of window should you use?
5.
Deploy ...................... when you need a NoSQL database of the supported API model, at planet scale, and with low latency performance.
6.
Before we can create an Azure Cosmos DB container with an analytical store, we must first enable
7.
You have an Azure Synapse Analytics dedicated SQL pool. You need to ensure that data in the pool is encrypted at rest. The solution must NOT require modifying applications that query the data. What should you do?
8.
In Azure Synapse Studio, Which Hub you can use to manage SQL scripts, Synapse notebooks, data flows, and Power BI reports
9.
Which slowly changing dimension types is described in the below image
10.
In Data Lake Storage Gen2, it's recommended to store your data as many small files
11.
Which slowly changing dimension types is described in the below image
12.
You have two Azure Data Factory instances named dev1 and prod1. dev1 connects to an Azure DevOps Git repository. You publish changes from the main branch of the Git repository to dev1. You need to deploy the artifacts from dev1 to prod1. What should you do first?
13.
...................... is a simple, cost-effective solution for managing and scaling multiple databases that have varying and unpredictable usage demands.
14.
You plan to create an Azure Synapse Analytics dedicated SQL pool. You need to minimize the time it takes to identify queries that return confidential information as defined by the company's data privacy regulations and the users who executed the queues. Which two components should you include in the solution?
15.
You are building an Azure Analytics query that will receive input data from Azure IoT Hub and write the results to Azure Blob storage. You need to calculate the difference in readings per sensor per hour. Which query should you use?
16.
.............................. is an orchestration of pipeline activities that includes chaining activities in a sequence, branching, defining parameters at the pipeline level, and passing arguments while invoking the pipeline on-demand or from a trigger
17.
Which Azure Synapse Analytics component enables you to perform Hybrid Transactional and Analytical Processing?
18.
Azure Data Lake Storage is a comprehensive, scalable, and cost-effective data lake solution for big data analytics built into Azure.
19.
In Azure Synapse Studio, the workspace tab of the data hub is where you can view the contents of the primary data lake store.
20.
............... is a column with a unique identifier for each row, and not generated from the table data, and the Data modelers like to create it on their tables when they design data warehouse models.
21.
.............. is the cloud-based ETL and data integration service that allows you to create data-driven workflows for orchestrating data movement and transforming data at scale.
22.
In Azure Synapse Studio, Which Hub you can use to access your provisioned SQL pool databases and SQL serverless databases in your workspace.
23.
AzCopy supports a maximum file size of .................. and automatically splits data files that exceed ...................
24.
Which statement is correct about Vertical Partitioning?
25.
.............. is a high-performance, column-oriented format optimized for big data scenarios.
26.
Spark pools can be shut down with no loss of data
27.
The ....................... gives the workspace SQL engines access to databases and tables created with Spark.
28.
The staging area is a neutral storage area that sits between the source systems and the data warehouse
29.
You have an Azure Synapse Analytics dedicated SQL pool that contains a large fact table. The table contains 50 columns and 5 billion rows and is a heap. Most queries against the table aggregate values from approximately 100 million rows and return only two columns. You discover that the queries against the fact table are very slow. Which type of index should you add to provide the fastest query times?
30.
Azure Synapse Pipelines is based in the Azure Data Factory service.
31.
Which slowly changing dimension types is described in the below image
32.
Spark pools in Azure Synapse can only use BLOB storage
33.
Which slowly changing dimension types is described in the below image
34.
................. is a Hadoop-compatible data repository that can store any size or type of data.
35.
The .......... is the ideal storage solution for hosting staging data in Modern Data Warehouse
36.
The data engineering team manages Azure HDInsight clusters. The team spends a large amount of time creating and destroying clusters daily because most of the data pipeline process runs in minutes. You need to implement a solution that deploys multiple HDInsight clusters with minimal effort. What should you implement?
37.
Use ..................... when you need to scale up and scale down OLTP systems on demand.
38.
In Azure Synapse Studio, The workspace tab of the data hub is where you can view the contents of the SQL and Spark pools.
39.
............... is an open-source distributed system that is used for processing big data workloads
40.
You are designing an Azure SQL Database that will use elastic pools. You plan to store data about customers in a table. Each record uses a value for CustomerID. What's the recommended strategy to partition data based on values in CustomerID?
41.
Apache Spark processes large amounts of data in Hard Disk
42.
Columnstore tables generally won't push data into a compressed columnstore segment until there are more than .............. rows per table.
43.
...................... is designed for customers looking to migrate a large number of apps from on- premises or IaaS, self-built, or ISV provided environment to fully managed PaaS cloud environment, with as low migration effort as possible.
44.
Azure Synapse Pipelines has an ....................... that enables it to bridge between the activity and linked Services objects.
Are you sure, you would like to submit your responses on DP-203 Exam Questions and Answers (Part 2) and view your results?