DP-200 Exam QuestionsBrowse all questions from this exam

DP-200 Exam - Question 202


A company plans to store hundreds of files in an Azure Storage account and in Azure Data Lake Storage account. The files will be stored in the parquet format. A solution must be in place that would adopt the following requirements:

- Provide the ability to process the data every 5 hours

- Give the ability for interactive data analysis

- Give the ability to process data using solid-state drive caching

- Make use of Directed Acyclic Graph processing mechanisms

- Provide support for REST API calls for monitoring purposes

- Ensure support for Python and Integration with Microsoft Power BI

Which of the following would you consider for the solution?

Show Answer
Correct Answer: D

Given the requirements: the ability to process data every 5 hours, support for interactive data analysis, solid-state drive caching, Directed Acyclic Graph processing, support for REST API calls, and ensuring support for Python and integration with Microsoft Power BI, the best choice is an HDInsight Spark cluster. Spark provides processing capabilities and uses Directed Acyclic Graph (DAG) for job scheduling. It also supports Python (via PySpark), integrates well with Power BI, and offers SSD-based caching for performance improvements. Additionally, Spark clusters can be monitored using REST APIs. Therefore, an HDInsight Spark cluster fulfills all the specified requirements effectively.

Discussion

2 comments
Sign in to comment
Avinash75
Jul 11, 2021

should be D .. Spark Cluster has all the requirements as mentioned in the reference link : https://docs.microsoft.com/en-us/azure/hdinsight/spark/apache-spark-overview

massnonn
Nov 22, 2021

"Ensure support for Python " this is seems you have to use spark