1.
To do ad-hoc data transfers into an Azure Data Lake store, use .....................
2.
You have a C# application that process data from an Azure IoT hub and performs complex transformations. You need to replace the application with a real-time solution. The solution must reuse as much code as possible from the existing application.
3.
If you want to store data without performing analysis on the data, set up the storage account as an
4.
There are four stages for processing big data solutions that are common to all architectures, one of theses stages is ........................ that involves the technologies that will present the data to users.
5.
You are designing security for administrative access to Azure Synapse Analytics. You need to recommend a solution to ensure that administrators use two-factor authentication when accessing the data warehouse from Microsoft SQL Server Management Studio (SSMS). What should you include in the recommendation?
6.
You have an Azure Databricks workspace named workspace1 in the Standard pricing tier. Workspace1 contains an all-purpose cluster named cluster1. You need to reduce the time it takes for cluster1 to start and scale up. The solution must minimize costs. What should you do first?
7.
You design data engineering solutions for a company. A project requires analytics and visualization of large set of data. The project has Notebook scheduling, Cluster automation, and Power BI Visualization You need to recommend the appropriate Azure service. Your solution must minimize the number of services required. Which Azure service should you recommend?
8.
One of Data Lake Storage Gen2 limitation is that you can't treat the data as if it's stored in a Hadoop Distributed File System
9.
......................... is a free application for Windows, macOS, and Linux, designed to manage unstructured data such as tables, blobs, queues, and files.
10.
.............. is a cloud-based data integration service that creates workflows in the cloud.
11.
You have a Windows-based solution that analyzes scientific data. You are designing a cloud-based solution that performs real-time analysis of the data. You need to design the logical flow for the solution. Which two actions should you recommend?
12.
A company stores sensitive information about customers and employees in Azure SQL Database. You need to ensure that the sensitive data remains encrypted in transit and at rest. What should you recommend?
13.
There are four stages for processing big data solutions that are common to all architectures, one of theses stages is ........................ that identifies the technology and processes that are used to acquire the source data.
14.
............................ uses a method that encrypts data in a less predictable manner. It is more secure, but prevents searching, grouping, indexing, and joining on encrypted columns.
15.
You need to implement an Azure Storage account that will use a Blob service endpoint that uses zone-redundant storage (ZRS). The storage account must only accept connections from a virtual network over Azure Private Link. What should you include in the implementation?
16.
You are planning an Azure solution that will aggregate streaming data. The input data will be retrieved from tab-separated values (TSV) files in Azure Blob storage. You need to output the maximum value from a specific column for every two-minute period in near real-time. The output must be written to Blob storage as a Parquet file. What should you use?
17.
A limitation of server clustering is that the hardware for each server in the cluster must be ................
18.
You are designing a solution that will use Azure Databricks and Azure Data Lake Storage Gen2. From Databricks, you need to access Data Lake Storage directly by using a service principal. What should you include in the solution?
19.
Your company has a Data Lake Storage Gen2 account. If you want to upload a single file by using a tool that you don't have to install or configure, which tool should you use?
20.
One of the Event processing components which generates an event data stream?
21.
If you want to copy small data volume from AWS S3 to Azure (for example, less than 10 TB), it's more efficient and easy to use the ....................
22.
You have an Azure Storage account. You plan to copy one million image files to the storage account. You plan to share the files with an external partner organization. The partner organization will analyze the files during the next year. You need to recommend an external access solution for the storage account. The solution must ensure that only the partner organization can access the storage account. Ensure that access of the partner organization is removed automatically after 365 days. What should you include in the recommendation?
23.
There are four stages for processing big data solutions that are common to all architectures, one of theses stages is ........................ that identifies the technologies that are used to perform data preparation and model training and scoring for data science solutions.
24.
If you want to enable the best performance for analytical workloads in Data Lake Storage Gen2, then on the Advanced tab of the Storage Account creation set the Hierarchical Namespace to ..................
25.
You have a large amount of sensor data stored in an Azure Data Lake Storage Gen2 account. The files are in the Parquet file format. New sensor data will be published to Azure Event Hubs. You need to recommend a solution to add the new sensor data to the existing sensor data in real-time. The solution must support the interactive querying of the entire dataset. Which type of server should you include in the recommendation?
26.
You are planning a big data solution in Azure. You need to recommend a technology that meets the following requirements: Be optimized for batch processing, Support autoscaling, and Support per-cluster scaling. Which technology should you recommend?
27.
Lift and shift an application from a physical environment to Azure Virtual Machines requires re-architecting the application.
28.
In Event processing, What is the application that consumes the output of an event processor?
29.
Which Azure Storage option is NoSQL store for no-schema storage of structured data?
30.
You are designing an Azure Synapse solution that will provide a query interface for the data stored in an Azure Storage account. The storage account is only accessible from a virtual network. You need to recommend an authentication mechanism to ensure that the solution can access the source data. What should you recommend?
31.
In Event processing, What's the engine that designed to consume and transform event data streams?
32.
When moving to the cloud, many customers migrate from physical or virtualized on-premises servers to Azure Virtual Machines. This strategy is known as ..................
33.
You are designing a solution that will copy Parquet files stored in an Azure Blob storage account to an Azure Data Lake Storage Gen2 account. The data will be loaded daily to the data lake and will use a folder structure of {Year}/{Month}/{Day}/. You need to design a daily Azure Data Factory data load to minimize the data transfer between the two accounts. Which two configurations should you include in the design?
34.
.......................... is a fully managed cloud service that records information about data stores.
35.
Which phase the Azure Data Lake Storage resides for processing big data solution?
36.
You have an Azure SQL database that has columns. The columns contain sensitive Personally Identifiable Information (PII) data. You need to design a solution that tracks and stores all the queries executed against the PII data. You must be able to review the data in Azure Monitor, and the data must be available for at least 45 days. Which solution should you do?
37.
You are developing an application that uses Azure Data Lake Storage Gen 2. You need to recommend a solution to grant permissions to a specific application for a limited time period. What should you include in the recommendation?
38.
There are four stages for processing big data solutions that are common to all architectures, one of theses stages is ........................ that identifies where the ingested data should be placed. In this case, we're using Azure Data Lake Storage Gen2.
39.
............................ is the quickest way to get event processing running on Azure where you can ingest streaming data from Azure Events Hubs or IoT Hub and run real-time analytics queries against the streams.
40.
If you are performing analytics on the data, set up the storage account as an
41.
You are designing a storage solution to store CSV files. You need to grant a data scientist access to read all the files in a single container of an Azure Storage account. The solution must use the principle of least privilege and provide the highest level of security. What are two possible ways to achieve the goal?
42.
You are designing a real-time stream processing solution in Azure Stream Analytics. The solution must read data from a blob container in an Azure Storage account via a service endpoint. You need to recommend an authentication mechanism for the solution. What should you recommend?
43.
You can enable a hierarchical namespace on an existing storage account.
44.
You are planning a solution to aggregate streaming data that originates in Apache Kafka and is output to Azure Data Lake Storage Gen2. The developers who will implement the stream processing solution use Java. Which service should you recommend using to process the streaming data?
Are you sure, you would like to submit your responses on DP-203 Exam Questions and Answers (Part 8) and view your results?