Question 121

You are designing a monitoring solution for a fleet of 500 vehicles. Each vehicle has a GPS tracking device that sends data to an Azure event hub once per minute.
You have a CSV file in an Azure Data Lake Storage Gen2 container. The file maintains the expected geographical area in which each vehicle should be.
You need to ensure that when a GPS position is outside the expected area, a message is added to another event hub for processing within 30 seconds. The solution must minimize cost.
What should you include in the solution? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.

Question 122

You have data stored in thousands of CSV files in Azure Data Lake Storage Gen2. Each file has a header row followed by a properly formatted carriage return (/r) and line feed (/n).
You are implementing a pattern that batch loads the files daily into an enterprise data warehouse in Azure Synapse Analytics by using PolyBase.
You need to skip the header row when you import the files into the data warehouse. Before building the loading pattern, you need to prepare the required database objects in Azure Synapse Analytics.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
NOTE: Each correct selection is worth one point

Question 123

You have an Azure Data lake Storage account that contains a staging zone.
You need to design a daily process to ingest incremental data from the staging zone, transform the data by executing an R script, and then insert the transformed data into a data warehouse in Azure Synapse Analytics.
Solution You use an Azure Data Factory schedule trigger to execute a pipeline that executes an Azure Databricks notebook, and then inserts the data into the data warehouse Dow this meet the goal?
  • Question 124

    You have an Azure Synapse workspace named MyWorkspace that contains an Apache Spark database named mytestdb.
    You run the following command in an Azure Synapse Analytics Spark pool in MyWorkspace.
    CREATE TABLE mytestdb.myParquetTable(
    EmployeeID int,
    EmployeeName string,
    EmployeeStartDate date)
    USING Parquet
    You then use Spark to insert a row into mytestdb.myParquetTable. The row contains the following data.

    One minute later, you execute the following query from a serverless SQL pool in MyWorkspace.
    SELECT EmployeeID
    FROM mytestdb.dbo.myParquetTable
    WHERE name = 'Alice';
    What will be returned by the query?
  • Question 125

    You are designing an Azure Synapse Analytics dedicated SQL pool.
    You need to ensure that you can audit access to Personally Identifiable information (PII).
    What should you include in the solution?