Online Access Free 70-776 Practice Test
| Exam Code: | 70-776 |
| Exam Name: | Engineering Data with Microsoft Cloud Services |
| Certification Provider: | Microsoft |
| Free Question Number: | 92 |
| Posted: | Oct 22, 2025 |
You plan to create for an alert for a Microsoft Azure Data Factory pipeline.
You need to configure the alert to trigger when the total number of failed runs exceeds five within a three-hour period.
How should you configure the window size and the threshold in the JSON file? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
Note: This question is part of a series of questions that present the same scenario. For your convenience, the scenario is repeated in each question. Each question presents a different goal and answer choices, but the text of the scenario is exactly the same in each question in this series.
Start of repeated scenario
You are migrating an existing on-premises data warehouse named LocalDW to Microsoft Azure. You will use an Azure SQL data warehouse named AzureDW for data storage and an Azure Data Factory named AzureDF for extract, transformation, and load (ETL) functions.
For each table in LocalDW, you create a table in AzureDW.
On the on-premises network, you have a Data Management Gateway.
Some source data is stored in Azure Blob storage. Some source data is stored on an on-premises Microsoft SQL Server instance. The instance has a table named Table1.
After data is processed by using AzureDF, the data must be archived and accessible forever. The archived data must meet a Service Level Agreement (SLA) for availability of 99 percent. If an Azure region fails, the archived data must be available for reading always.
End of repeated scenario.
You need to configure Azure Data Factory to connect to the on-premises SQL Server instance.
What should you do first?
You have an on-premises data warehouse that uses Microsoft SQL Server 2016. All the data in the data warehouse comes from text files stored in Azure Blob storage. The text files are imported into the data warehouse by using SQL Server Integration Services (SSIS). The text files are not transformed.
You need to migrate the data to an Azure SQL data warehouse in the least amount of time possible.
Which two actions should you perform? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
You have a Microsoft Azure SQL data warehouse. The following statements are used to define file formats in the data warehouse.
You have an external PolyBase table named file_factPowerMeasurement that uses the FileFormat_ORC file format.
You need to change file_ factPowerMeasurement to use the FileFormat_PARQUET file format.
Which two statements should you execute? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this sections, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You are monitoring user queries to a Microsoft Azure SQL data warehouse that has six compute nodes.
You discover that compute node utilization is uneven. The rows_processed column from sys.dm_pdw_workers shows a significant variation in the number of rows being moved among the distributions for the same table for the same query.
You need to ensure that the load is distributed evenly across the compute nodes.
Solution: You change the table to use a column that is not skewed for hash distribution.
Does this meet the goal?

