What should you do?

Your team is building a data engineering and data science development environment.

The environment must support the following requirements:

– support Python and Scala

– compose data storage, movement, and processing services into automated data pipelines

– the same tool should be used for the orchestration of both data engineering and data science

– support workload isolation and interactive workloads

– enable scaling across a cluster of machines

You need to create the environment.

What should you do?
A . Build the environment in Apache Hive for HDInsight and use Azure Data Factory for orchestration.
B . Build the environment in Azure Databricks and use Azure Data Factory for orchestration.
C . Build the environment in Apache Spark for HDInsight and use Azure Container Instances for orchestration.
D . Build the environment in Azure Databricks and use Azure Container Instances for orchestration.

Answer: B

Explanation:

In Azure Databricks, we can create two different types of clusters.

– Standard, these are the default clusters and can be used with Python, R, Scala and SQL

– High-concurrency

Azure Databricks is fully integrated with Azure Data Factory.

Incorrect Answers:

D: Azure Container Instances is good for development or testing. Not suitable for production workloads.

References: https://docs.microsoft.com/en-us/azure/architecture/data-guide/technology-choices/data-science-and­machine-learning

Latest DP-100 Dumps Valid Version with 227 Q&As

Latest And Valid Q&A | Instant Download | Once Fail, Full Refund

Subscribe
Notify of
guest
0 Comments
Inline Feedbacks
View all comments