HomeMicrosoft Exams PracticeDP-203 Exam Questions and Answers (P2) DP-203 Exam Questions and Answers (P2) Welcome to your DP-203 Exam Questions and Answers (Part 2) Please, provide your Full Name and Email to get started! Please, enter your Full Name Please, enter your Email 1. Which slowly changing dimension types is described in the below image Type 6 SCDType 1 SCDType 2 SCDType 3 SCD2. In Azure Synapse Studio, Which Hub you can use to access your provisioned SQL pool databases and SQL serverless databases in your workspace.Develop hubHome hubMonitor hubIntegrate hubManage hubData hub3. In Azure Synapse Studio, Which Hub you can use to manage SQL scripts, Synapse notebooks, data flows, and Power BI reportsHome hubMonitor hubIntegrate hubManage hubDevelop hubData hub4. You need to design an Azure Synapse Analytics dedicated SQL pool that can return an employee record from a given point in time, maintains the latest employee information, and minimizes query complexity. How should you model the employee data?A temporary tableA temporal tableA Type 2 slowly changing dimension table5. Deploy ...................... when you need a NoSQL database of the supported API model, at planet scale, and with low latency performance.Azure SQL DatabaseAzure Cosmos DatabaseAzure Data Lake6. You are building an Azure Analytics query that will receive input data from Azure IoT Hub and write the results to Azure Blob storage. You need to calculate the difference in readings per sensor per hour. Which query should you use?SELECT sensorId, growth = reading - LEAD(reading) OVER (PARTITION BY sensorId LIMIT DURATION(hour, 1)) FROM inputSELECT sensorId, growth = reading - LAST(reading) OVER (PARTITION BY sensorId WHEN(hour, 1)) FROM inputSELECT sensorId, growth = reading - LAG(reading) OVER (PARTITION BY sensorId LIMIT DURATION(hour, 1)) FROM input7. The data engineering team manages Azure HDInsight clusters. The team spends a large amount of time creating and destroying clusters daily because most of the data pipeline process runs in minutes. You need to implement a solution that deploys multiple HDInsight clusters with minimal effort. What should you implement?Azure Resource Manager templatesAzure Traffic ManagerAzure Databricks8. You plan to create an Azure Synapse Analytics dedicated SQL pool. You need to minimize the time it takes to identify queries that return confidential information as defined by the company's data privacy regulations and the users who executed the queues. Which two components should you include in the solution? audit logs sent to a Log Analytics workspace dynamic data masking for columns that contain confidential information sensitivity-classification labels applied to columns that contain confidential information resource tags for databases that contain confidential information9. In Data Lake Storage Gen2, it's recommended to store your data as many small filesFalseTrue10. .............. is a high-performance, column-oriented format optimized for big data scenarios.CSVORCTXTParquetJSON11. ................. is a Hadoop-compatible data repository that can store any size or type of data.Queue StorageAzure Data Lake StorageTable StorageBlob Storage12. You have an Azure Synapse Analytics dedicated SQL pool that contains a large fact table. The table contains 50 columns and 5 billion rows and is a heap. Most queries against the table aggregate values from approximately 100 million rows and return only two columns. You discover that the queries against the fact table are very slow. Which type of index should you add to provide the fastest query times?nonclusteredclusteredclustered columnstorenonclustered columnstore13. In Azure Synapse Studio, The workspace tab of the data hub is where you can view the contents of the SQL and Spark pools.TrueFalse14. Which statement is correct about Vertical Partitioning?The data is partitioned between a number of databases in a data tier. Typically, different sets of tables reside on different databases.The data is partitioned to distribute rows across a scaled out data tier. With this approach, the schema is identical on all participating databases.15. AzCopy supports a maximum file size of .................. and automatically splits data files that exceed ...................1 TB , 200 GB10 TB , 200 GB1 GB , 200 MB16. The ....................... gives the workspace SQL engines access to databases and tables created with Spark.Azure Synapse Spark pools.Azure Synapse SQL.Azure Synapse Link.Azure Synapse shared metadata.17. You have two Azure Data Factory instances named dev1 and prod1. dev1 connects to an Azure DevOps Git repository. You publish changes from the main branch of the Git repository to dev1. You need to deploy the artifacts from dev1 to prod1. What should you do first?From Azure DevOps, update the main branch.From dev1, modify the Git configuration.From Azure DevOps, create a release pipeline.From dev1, create a linked service.18. Which slowly changing dimension types is described in the below imageType 6 SCDType 3 SCDType 1 SCDType 2 SCD19. .............. is the cloud-based ETL and data integration service that allows you to create data-driven workflows for orchestrating data movement and transforming data at scale.Linked ServicesAzure Synapse PipelinesSpark PoolMapping Data Flow20. ...................... is designed for customers looking to migrate a large number of apps from on- premises or IaaS, self-built, or ISV provided environment to fully managed PaaS cloud environment, with as low migration effort as possible.Azure SQL Database elastic poolsAzure SQL Database Managed InstanceAzure SQL Database Hyperscale21. ............... is a column with a unique identifier for each row, and not generated from the table data, and the Data modelers like to create it on their tables when they design data warehouse models.A surrogate keyA foreign keyA business key22. Azure Synapse Pipelines is based in the Azure Data Factory service.TrueFalse23. Which slowly changing dimension types is described in the below imageType 1 SCDType 3 SCDType 2 SCDType 6 SCD24. Spark pools in Azure Synapse can only use BLOB storageTrueFalse25. Azure Data Lake Storage is a comprehensive, scalable, and cost-effective data lake solution for big data analytics built into Azure.TrueFalse26. You are designing an Azure Stream Analytics job to process incoming events from sensors in retail environments. You need to process the events to produce a running average of shopper counts during the previous 15 minutes, calculated at five-minute intervals. Which type of window should you use?Sliding WindowTumbling WindowSession WindowHopping Window27. Before we can create an Azure Cosmos DB container with an analytical store, we must first enableSpark PoolAzure Synapse LinkPipelinesAzure Synapse SQL28. You are designing an Azure SQL Database that will use elastic pools. You plan to store data about customers in a table. Each record uses a value for CustomerID. What's the recommended strategy to partition data based on values in CustomerID?Separate data into shards by using Horizontal partitioningSeparate data into customer ID by using Horizontal partitioningSeparate data into customer regions by using Horizontal partitioningSeparate data into customer regions by using Vertical partitioning29. You have an Azure Synapse Analytics dedicated SQL pool. You need to ensure that data in the pool is encrypted at rest. The solution must NOT require modifying applications that query the data. What should you do?Enable Transparent Data Encryption (TDE) for the pool.Create an Azure key vault in the Azure subscription grant access to the pool.Use a customer-managed key to enable double encryption for the Azure Synapse workspace.Enable encryption at rest for the Azure Data Lake Storage Gen2 account.30. Azure Synapse Pipelines has an ....................... that enables it to bridge between the activity and linked Services objects.ArgumentsParametersControl flowIntegration runtime31. ...................... is a simple, cost-effective solution for managing and scaling multiple databases that have varying and unpredictable usage demands.Azure SQL Database HyperscaleAzure SQL Database elastic poolsAzure SQL Database Managed Instance32. ...................... supports up to 100 TB of data and provides high throughput and performance, as well as rapid scaling to adapt to the workload requirements.Azure SQL Database Managed InstanceAzure SQL Database HyperscaleAzure SQL Database elastic pools33. .............................. is an orchestration of pipeline activities that includes chaining activities in a sequence, branching, defining parameters at the pipeline level, and passing arguments while invoking the pipeline on-demand or from a triggerArgumentsControl flowIntegration runtimeParameters34. Which slowly changing dimension types is described in the below imageType 3 SCDType 1 SCDType 2 SCDType 6 SCD35. Apache Spark processes large amounts of data in Hard DiskFalseTrue36. ............... is an open-source distributed system that is used for processing big data workloadsApache SparkAzure Synapse ServiceDate LakeAzure Data Factory37. Use ..................... when you need to scale up and scale down OLTP systems on demand.Azure SQL DatabaseAzure Data LakeAzure Cosmos Database38. Which Azure Synapse Analytics component enables you to perform Hybrid Transactional and Analytical Processing?Azure Synapse Pipeline.Azure Synapse SQL.Azure Synapse Studio.Azure Synapse Link.39. Columnstore tables generally won't push data into a compressed columnstore segment until there are more than .............. rows per table.1 million6 million60 million10 million40. In Azure Synapse Studio, the workspace tab of the data hub is where you can view the contents of the primary data lake store.FalseTrue41. Which statement is correct about Horizontal Partitioning?The data is partitioned to distribute rows across a scaled out data tier. With this approach, the schema is identical on all participating databases.The data is partitioned between a number of databases in a data tier. Typically, different sets of tables reside on different databases.42. The staging area is a neutral storage area that sits between the source systems and the data warehouseTrueFalse43. The .......... is the ideal storage solution for hosting staging data in Modern Data WarehouseBlob storageAzure FilesAzure Data Lake Gen 2Azure Synapse Spark Pools44. Spark pools can be shut down with no loss of dataTrueFalse12 out of 11Are you sure, you would like to submit your responses on DP-203 Exam Questions and Answers (Part 2) and view your results? Time is Up! Related