1.
Apache Spark processes large amounts of data in Hard Disk
2.
...................... is designed for customers looking to migrate a large number of apps from on- premises or IaaS, self-built, or ISV provided environment to fully managed PaaS cloud environment, with as low migration effort as possible.
3.
You have an Azure Synapse Analytics dedicated SQL pool. You need to ensure that data in the pool is encrypted at rest. The solution must NOT require modifying applications that query the data. What should you do?
4.
You plan to create an Azure Synapse Analytics dedicated SQL pool. You need to minimize the time it takes to identify queries that return confidential information as defined by the company's data privacy regulations and the users who executed the queues. Which two components should you include in the solution?
5.
Use ..................... when you need to scale up and scale down OLTP systems on demand.
6.
Which slowly changing dimension types is described in the below image
7.
Azure Synapse Pipelines is based in the Azure Data Factory service.
8.
...................... is a simple, cost-effective solution for managing and scaling multiple databases that have varying and unpredictable usage demands.
9.
In Data Lake Storage Gen2, it's recommended to store your data as many small files
10.
The data engineering team manages Azure HDInsight clusters. The team spends a large amount of time creating and destroying clusters daily because most of the data pipeline process runs in minutes. You need to implement a solution that deploys multiple HDInsight clusters with minimal effort. What should you implement?
11.
AzCopy supports a maximum file size of .................. and automatically splits data files that exceed ...................
12.
You need to design an Azure Synapse Analytics dedicated SQL pool that can return an employee record from a given point in time, maintains the latest employee information, and minimizes query complexity. How should you model the employee data?
13.
Which slowly changing dimension types is described in the below image
14.
Before we can create an Azure Cosmos DB container with an analytical store, we must first enable
15.
Which slowly changing dimension types is described in the below image
16.
Spark pools in Azure Synapse can only use BLOB storage
17.
.............. is a high-performance, column-oriented format optimized for big data scenarios.
18.
Spark pools can be shut down with no loss of data
19.
In Azure Synapse Studio, Which Hub you can use to manage SQL scripts, Synapse notebooks, data flows, and Power BI reports
20.
The .......... is the ideal storage solution for hosting staging data in Modern Data Warehouse
21.
.............. is the cloud-based ETL and data integration service that allows you to create data-driven workflows for orchestrating data movement and transforming data at scale.
22.
Deploy ...................... when you need a NoSQL database of the supported API model, at planet scale, and with low latency performance.
23.
In Azure Synapse Studio, Which Hub you can use to access your provisioned SQL pool databases and SQL serverless databases in your workspace.
24.
Which Azure Synapse Analytics component enables you to perform Hybrid Transactional and Analytical Processing?
25.
Azure Synapse Pipelines has an ....................... that enables it to bridge between the activity and linked Services objects.
26.
You are designing an Azure SQL Database that will use elastic pools. You plan to store data about customers in a table. Each record uses a value for CustomerID. What's the recommended strategy to partition data based on values in CustomerID?
27.
You have two Azure Data Factory instances named dev1 and prod1. dev1 connects to an Azure DevOps Git repository. You publish changes from the main branch of the Git repository to dev1. You need to deploy the artifacts from dev1 to prod1. What should you do first?
28.
Which slowly changing dimension types is described in the below image
29.
You are building an Azure Analytics query that will receive input data from Azure IoT Hub and write the results to Azure Blob storage. You need to calculate the difference in readings per sensor per hour. Which query should you use?
30.
............... is an open-source distributed system that is used for processing big data workloads
31.
Columnstore tables generally won't push data into a compressed columnstore segment until there are more than .............. rows per table.
32.
Which statement is correct about Vertical Partitioning?
33.
You have an Azure Synapse Analytics dedicated SQL pool that contains a large fact table. The table contains 50 columns and 5 billion rows and is a heap. Most queries against the table aggregate values from approximately 100 million rows and return only two columns. You discover that the queries against the fact table are very slow. Which type of index should you add to provide the fastest query times?
34.
Which statement is correct about Horizontal Partitioning?
35.
...................... supports up to 100 TB of data and provides high throughput and performance, as well as rapid scaling to adapt to the workload requirements.
36.
In Azure Synapse Studio, the workspace tab of the data hub is where you can view the contents of the primary data lake store.
37.
............... is a column with a unique identifier for each row, and not generated from the table data, and the Data modelers like to create it on their tables when they design data warehouse models.
38.
................. is a Hadoop-compatible data repository that can store any size or type of data.
39.
In Azure Synapse Studio, The workspace tab of the data hub is where you can view the contents of the SQL and Spark pools.
40.
.............................. is an orchestration of pipeline activities that includes chaining activities in a sequence, branching, defining parameters at the pipeline level, and passing arguments while invoking the pipeline on-demand or from a trigger
41.
Azure Data Lake Storage is a comprehensive, scalable, and cost-effective data lake solution for big data analytics built into Azure.
42.
The ....................... gives the workspace SQL engines access to databases and tables created with Spark.
43.
You are designing an Azure Stream Analytics job to process incoming events from sensors in retail environments. You need to process the events to produce a running average of shopper counts during the previous 15 minutes, calculated at five-minute intervals. Which type of window should you use?
44.
The staging area is a neutral storage area that sits between the source systems and the data warehouse
Are you sure, you would like to submit your responses on DP-203 Exam Questions and Answers (Part 2) and view your results?