You are currently asked to work on building a data pipeline, you have noticed that you are currently working on a very large scale ETL many data dependencies, which of the following tools can be used to address this problem?
You are currently asked to work on building a data pipeline, you have noticed that you are currently working on a very large scale ETL many data dependencies, which of the following tools can be used to address this problem?A . AUTO LOADER B. JOBS and TASKS C. SQL Endpoints...
AS SELECT * FROM table_name
AS SELECT * FROM table_nameView AnswerAnswer: C Explanation:
You noticed that colleague is manually copying the notebook with _bkp to store the previous versions, which of the following feature would you recommend instead.
You noticed that colleague is manually copying the notebook with _bkp to store the previous versions, which of the following feature would you recommend instead.A . Databricks notebooks support change tracking and versioning B. Databricks notebooks should be copied to a local machine and setup source control locally to version...
You are currently asked to work on building a data pipeline, you have noticed that you are currently working on a very large scale ETL many data dependencies, which of the following tools can be used to address this problem?
You are currently asked to work on building a data pipeline, you have noticed that you are currently working on a very large scale ETL many data dependencies, which of the following tools can be used to address this problem?A . AUTO LOADER B. JOBS and TASKS C. SQL Endpoints...
Which of the following describes how Databricks Repos can help facilitate CI/CD workflows on the Databricks Lakehouse Platform?
Which of the following describes how Databricks Repos can help facilitate CI/CD workflows on the Databricks Lakehouse Platform?A . Databricks Repos can facilitate the pull request, review, and approval process before merging branches B. Databricks Repos can merge changes from a secondary Git branch into a main Git branch C....
table(table_name))
table(table_name))A . format, checkpointlocation, schemalocation, overwrite B. cloudfiles.format, checkpointlocation, cloudfiles.schemalocation, overwrite C. cloudfiles.format, cloudfiles.schemalocation, checkpointlocation, mergeSchema D. cloudfiles.format, cloudfiles.schemalocation, checkpointlocation, overwrite E. cloudfiles.format, cloudfiles.schemalocation, checkpointlocation, appendView AnswerAnswer: C Explanation: The answer is cloudfiles.format, cloudfiles.schemalocation, checkpointlocation, mergeSchema. Here is the end to end syntax of streaming ELT, below link contains complete...
What is the main difference between the silver layer and the gold layer in medallion architecture?
What is the main difference between the silver layer and the gold layer in medallion architecture?A . Silver may contain aggregated data B. Gold may contain aggregated data C. Data quality checks are applied in gold D. Silver is a copy of bronze data E. God is a copy of...
Which of the following techniques structured streaming uses to ensure recovery of failures during stream processing?
Which of the following techniques structured streaming uses to ensure recovery of failures during stream processing?A . Checkpointing and Watermarking B. Write ahead logging and watermarking C. Checkpointing and write-ahead logging D. Delta time travel E. The stream will failover to available nodes in the cluster F. Checkpointing and Idempotent...
What steps need to be taken to set up a DELTA LIVE PIPELINE as a job using the workspace UI?
What steps need to be taken to set up a DELTA LIVE PIPELINE as a job using the workspace UI?A . DELTA LIVE TABLES do not support job cluster B. Select Workflows UI and Delta live tables tab, under task type select Delta live tables pipeline and select the notebook...
Which of the following approaches can enable the data engineering team to be notified if the ELT job has not been run in an hour?
A data engineer is using a Databricks SQL query to monitor the performance of an ELT job. The ELT job is triggered by a specific number of input records being ready to process. The Databricks SQL query returns the number of minutes since the job’s most recent runtime. Which of...