1.
Azure Synapse Pipelines has an ....................... that enables it to bridge between the activity and linked Services objects.
2.
Columnstore tables generally won't push data into a compressed columnstore segment until there are more than .............. rows per table.
3.
Use ..................... when you need to scale up and scale down OLTP systems on demand.
4.
Which statement is correct about Horizontal Partitioning?
5.
In Data Lake Storage Gen2, it's recommended to store your data as many small files
6.
In Azure Synapse Studio, The workspace tab of the data hub is where you can view the contents of the SQL and Spark pools.
7.
Deploy ...................... when you need a NoSQL database of the supported API model, at planet scale, and with low latency performance.
8.
...................... supports up to 100 TB of data and provides high throughput and performance, as well as rapid scaling to adapt to the workload requirements.
9.
.............. is the cloud-based ETL and data integration service that allows you to create data-driven workflows for orchestrating data movement and transforming data at scale.
10.
You have an Azure Synapse Analytics dedicated SQL pool. You need to ensure that data in the pool is encrypted at rest. The solution must NOT require modifying applications that query the data. What should you do?
11.
............... is an open-source distributed system that is used for processing big data workloads
12.
Which slowly changing dimension types is described in the below image
13.
You have an Azure Synapse Analytics dedicated SQL pool that contains a large fact table. The table contains 50 columns and 5 billion rows and is a heap. Most queries against the table aggregate values from approximately 100 million rows and return only two columns. You discover that the queries against the fact table are very slow. Which type of index should you add to provide the fastest query times?
14.
............... is a column with a unique identifier for each row, and not generated from the table data, and the Data modelers like to create it on their tables when they design data warehouse models.
15.
Which slowly changing dimension types is described in the below image
16.
.............. is a high-performance, column-oriented format optimized for big data scenarios.
17.
Apache Spark processes large amounts of data in Hard Disk
18.
You plan to create an Azure Synapse Analytics dedicated SQL pool. You need to minimize the time it takes to identify queries that return confidential information as defined by the company's data privacy regulations and the users who executed the queues. Which two components should you include in the solution?
19.
AzCopy supports a maximum file size of .................. and automatically splits data files that exceed ...................
20.
Azure Synapse Pipelines is based in the Azure Data Factory service.
21.
Which statement is correct about Vertical Partitioning?
22.
In Azure Synapse Studio, the workspace tab of the data hub is where you can view the contents of the primary data lake store.
23.
Spark pools can be shut down with no loss of data
24.
Spark pools in Azure Synapse can only use BLOB storage
25.
Which slowly changing dimension types is described in the below image
26.
You have two Azure Data Factory instances named dev1 and prod1. dev1 connects to an Azure DevOps Git repository. You publish changes from the main branch of the Git repository to dev1. You need to deploy the artifacts from dev1 to prod1. What should you do first?
27.
Azure Data Lake Storage is a comprehensive, scalable, and cost-effective data lake solution for big data analytics built into Azure.
28.
.............................. is an orchestration of pipeline activities that includes chaining activities in a sequence, branching, defining parameters at the pipeline level, and passing arguments while invoking the pipeline on-demand or from a trigger
29.
................. is a Hadoop-compatible data repository that can store any size or type of data.
30.
In Azure Synapse Studio, Which Hub you can use to manage SQL scripts, Synapse notebooks, data flows, and Power BI reports
31.
In Azure Synapse Studio, Which Hub you can use to access your provisioned SQL pool databases and SQL serverless databases in your workspace.
32.
...................... is a simple, cost-effective solution for managing and scaling multiple databases that have varying and unpredictable usage demands.
33.
The .......... is the ideal storage solution for hosting staging data in Modern Data Warehouse
34.
Which slowly changing dimension types is described in the below image
35.
Which Azure Synapse Analytics component enables you to perform Hybrid Transactional and Analytical Processing?
36.
The staging area is a neutral storage area that sits between the source systems and the data warehouse
37.
You are designing an Azure SQL Database that will use elastic pools. You plan to store data about customers in a table. Each record uses a value for CustomerID. What's the recommended strategy to partition data based on values in CustomerID?
38.
Before we can create an Azure Cosmos DB container with an analytical store, we must first enable
39.
The data engineering team manages Azure HDInsight clusters. The team spends a large amount of time creating and destroying clusters daily because most of the data pipeline process runs in minutes. You need to implement a solution that deploys multiple HDInsight clusters with minimal effort. What should you implement?
40.
The ....................... gives the workspace SQL engines access to databases and tables created with Spark.
41.
You need to design an Azure Synapse Analytics dedicated SQL pool that can return an employee record from a given point in time, maintains the latest employee information, and minimizes query complexity. How should you model the employee data?
42.
You are building an Azure Analytics query that will receive input data from Azure IoT Hub and write the results to Azure Blob storage. You need to calculate the difference in readings per sensor per hour. Which query should you use?
43.
...................... is designed for customers looking to migrate a large number of apps from on- premises or IaaS, self-built, or ISV provided environment to fully managed PaaS cloud environment, with as low migration effort as possible.
44.
You are designing an Azure Stream Analytics job to process incoming events from sensors in retail environments. You need to process the events to produce a running average of shopper counts during the previous 15 minutes, calculated at five-minute intervals. Which type of window should you use?
Are you sure, you would like to submit your responses on DP-203 Exam Questions and Answers (Part 2) and view your results?