Certified Data Engineer Associate Exam QuestionsBrowse all questions from this exam

Certified Data Engineer Associate Exam - Question 75


A data engineer has joined an existing project and they see the following query in the project repository:

CREATE STREAMING LIVE TABLE loyal_customers AS

SELECT customer_id -

FROM STREAM(LIVE.customers)

WHERE loyalty_level = 'high';

Which of the following describes why the STREAM function is included in the query?

Show Answer
Correct Answer: C

The STREAM function is included in the query because the customers table is a streaming live table. This function is necessary to handle and process streaming data as it arrives in real-time, which is suitable for the context of a 'live table' that constantly updates based on incoming data streams.

Discussion

6 comments
Sign in to comment
meow_akkOption: C
Oct 22, 2023

Ans C is correct : https://docs.databricks.com/en/sql/load-data-streaming-table.html Load data into a streaming table To create a streaming table from data in cloud object storage, paste the following into the query editor, and then click Run: SQL Copy to clipboardCopy /* Load data from a volume */ CREATE OR REFRESH STREAMING TABLE <table-name> AS SELECT * FROM STREAM read_files('/Volumes/<catalog>/<schema>/<volume>/<path>/<folder>') /* Load data from an external location */ CREATE OR REFRESH STREAMING TABLE <table-name> AS SELECT * FROM STREAM read_files('s3://<bucket>/<path>/<folder>')

cxw23Option: A
Jan 3, 2024

Ans is A. CREATE STREAMING LIVE TABLE syntax is does not exist. It should be CREATE LIVE TABLE AS SELECT * FROM STREAM.

bartfto
Jan 10, 2024

LIVE references schema name customer_table references table name

azure_bimonsterOption: C
Jan 20, 2024

C is correct

OfficeSaracusOption: D
Apr 27, 2024

Option E, specifying "at least one notebook library to be executed," is not a requirement for setting up a Delta Live Tables pipeline. Delta Live Tables are built on top of Databricks and use notebooks to define the pipeline's logic, but the actual requirement when setting up the pipeline is typically the location where the data will be written to, like a target database or a path to cloud storage. While notebooks may contain the business logic for the transformations and actions within the pipeline, the fundamental requirement for setting up a pipeline is knowing where the data will reside after processing, hence why the location of the target database for the written data is crucial.

THC1138
May 25, 2024

Wrong question, that's for #73

THC1138
May 25, 2024

I mean question #74

benni_aleOption: C
Apr 29, 2024

c is ok

benni_aleOption: C
Apr 29, 2024

c is correct . about D: it can be correct but it is not given the fact it comes from pyspark ; sql supports (at least in databricks) the creation of streaming live table as well so it is not necessasarily from pyspark