Question 21

You plan to create an Azure Synapse Analytics dedicated SQL pool.
You need to minimize the time it takes to identify queries that return confidential information as defined by the company's data privacy regulations and the users who executed the queues.
Which two components should you include in the solution? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
  • Question 22

    You are designing an Azure Synapse Analytics dedicated SQL pool.
    Groups will have access to sensitive data in the pool as shown in the following table.

    You have policies for the sensitive dat
    a. The policies vary be region as shown in the following table.

    You have a table of patients for each region. The tables contain the following potentially sensitive columns.

    You are designing dynamic data masking to maintain compliance.
    For each of the following statements, select Yes if the statement is true. Otherwise, select No.
    NOTE: Each correct selection is worth one point.

    Question 23

    You have the following Azure Data Factory pipelines
    * ingest Data from System 1
    * Ingest Data from System2
    * Populate Dimensions
    * Populate facts
    ingest Data from System1 and Ingest Data from System1 have no dependencies. Populate Dimensions must execute after Ingest Data from System1 and Ingest Data from System* Populate Facts must execute after the Populate Dimensions pipeline. All the pipelines must execute every eight hours.
    What should you do to schedule the pipelines for execution?
  • Question 24

    Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
    After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
    You plan to create an Azure Databricks workspace that has a tiered structure. The workspace will contain the following three workloads:
    A workload for data engineers who will use Python and SQL.
    A workload for jobs that will run notebooks that use Python, Scala, and SOL.
    A workload that data scientists will use to perform ad hoc analysis in Scala and R.
    The enterprise architecture team at your company identifies the following standards for Databricks environments:
    The data engineers must share a cluster.
    The job cluster will be managed by using a request process whereby data scientists and data engineers provide packaged notebooks for deployment to the cluster.
    All the data scientists must be assigned their own cluster that terminates automatically after 120 minutes of inactivity. Currently, there are three data scientists.
    You need to create the Databricks clusters for the workloads.
    Solution: You create a Standard cluster for each data scientist, a High Concurrency cluster for the data engineers, and a Standard cluster for the jobs.
    Does this meet the goal?
  • Question 25

    Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
    After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
    You are designing an Azure Stream Analytics solution that will analyze Twitter dat a.
    You need to count the tweets in each 10-second window. The solution must ensure that each tweet is counted only once.
    Solution: You use a hopping window that uses a hop size of 10 seconds and a window size of 10 seconds.
    Does this meet the goal?