Question 6 of 83
HOTSPOT -
Note: This question is part of a series of questions that use the same scenario. For your convenience, the scenario is repeated in each question. Each question presents a different goal and answer choices, but the text of the scenario is exactly the same in each question in this series.

Start of repeated scenario -
You are developing a Microsoft Azure SQL data warehouse to perform analytics on the transit system of a city. The data warehouse will contain data about customers, trips, and community events.
You have two storage accounts named StorageAccount1 and StorageAccount2. StorageAccount1 is associated to the data warehouse. StorageAccount2 contains weather data files stored in the CSV format. The files have a naming format of city_state_yyymmdd.csv.
Microsoft SQL Server is installed on an Azure virtual machine named AzureVM1.
You are migrating from an existing on premises solution that uses Microsoft SQL Server 2016 Enterprise. The planned schema is shown in the exhibit. (Click the
Exhibit button)
Exam 70-776: Question 6 - Image 1
The first column of each table will contain unique values. A table named Customer will contain 12 million rows. A table named Trip will contain 3 billion rows.
You have the following view.
Exam 70-776: Question 6 - Image 2
You plan to use Azure Data Factory to perform the following four activities:
✑ Activity1: Invoke an R script to generate a prediction column.
✑ Activity2: Import weather data from a set of CSV files in Azure Blob storage
✑ Activity3: Execute a stored procedure in the Azure SQL data warehouse.
✑ Activity4: Copy data from an Amazon Simple Storage Service (S3).
You plan to detect the following two threat patterns:
✑ Pattern1: A user logs in from two physical locations.
✑ Pattern2: A user attempts to gain elevated permissions.

End of repeated scenario -
You plan to create a report that will query customer records for a selected ResidenceZip. The report will return customer trips sorted by TripStartDateTime.
You need to specify the distribution clause for each table. The solution must meet the following requirements.
✑ Minimize how long it takes to query the customer information.
✑ Perform the operation as a pass-through query without data movement.
How should you complete the statement? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
Exam 70-776: Question 6 - Image 3
    Correct Answer:
    The most common example of when a table distributed by a column will far outperform a Round Robin table is when two large fact tables are joined. For example, if you have an orders table, which is distributed by order_id, and a transactions table, which is also distributed by order_id, when you join your orders table to your transactions table on order_id, this query becomes a pass-through query, which means we eliminate data movement operations. Fewer steps mean a faster query. Less data movement also makes for faster queries.
    Incorrect Answers:
    Round Robin: Hash distribute large tables
    By default, tables are Round Robin distributed. This makes it easy for users to get started creating tables without having to decide how their tables should be distributed. Round Robin tables may perform sufficiently for some workloads, but in most cases selecting a distribution column will perform much better
    References: https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-best-practices Exam 70-776: Question 6 - Image 4
Question 7 of 83
You are designing a data loading process for a Microsoft Azure SQL data warehouse. Data will be loaded to Azure Blob storage, and then the data will be loaded to the data warehouse.
Which tool should you use to load the data to Azure Blob storage?
    Correct Answer: D

Question 8 of 83
You have a Microsoft Azure SQL data warehouse to which 1,000 Data Warehouse Units (DWUs) are allocated.
You plan to load 10 million rows of data to the data warehouse.
You need to load the data in the least amount of time possible. The solution must ensure that queries against the new data execute as quickly as possible.
What should you use to optimize the data load?
    Correct Answer: A

Question 9 of 83
HOTSPOT -
You need to create a Microsoft Azure SQL data warehouse named dw1 that supports up to 10 TB of data.
How should you complete the statement? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
Exam 70-776: Question 9 - Image 1
    Correct Answer:
    Box 1: distributed_size =10240 -
    10240 GB equals 10 TB.
    Parameter: distributed_size [ GB ]
    A positive number. The size, in integer or decimal gigabytes, for the total space allocated to distributed tables (and corresponding data) across the appliance.

    Box 2: datawarehouse -
    The EDITION parameter specifies the service tier of the database. For SQL Data Warehouse use 'datawarehouse' .
    References: https://docs.microsoft.com/en-us/sql/t-sql/statements/create-database-azure-sql-data-warehouse Exam 70-776: Question 9 - Image 2
Question 10 of 83
You have a Microsoft Azure SQL data warehouse stored in geo-redundant storage.
You experience a regional outage.
You plan to recover the database to a new region.
You need to get a list of the backup files that can be restored to the new region.
Which cmdlet should you run?
    Correct Answer: B