Databricks Certified Data Engineer Professional Exam Databricks-Certified-Professional-Data-Engineer Question # 27 Topic 3 Discussion

Databricks Certified Data Engineer Professional Exam Databricks-Certified-Professional-Data-Engineer Question # 27 Topic 3 Discussion

Databricks-Certified-Professional-Data-Engineer Exam Topic 3 Question 27 Discussion:
Question #: 27
Topic #: 3

In order to prevent accidental commits to production data, a senior data engineer has instituted a policy that all development work will reference clones of Delta Lake tables. After testing both deep and shallow clone, development tables are created using shallow clone.

A few weeks after initial table creation, the cloned versions of several tables implemented as Type 1 Slowly Changing Dimension (SCD) stop working. The transaction logs for the source tables show that vacuum was run the day before.

Why are the cloned tables no longer working?


A.

The data files compacted by vacuum are not tracked by the cloned metadata; running refresh on the cloned table will pull in recent changes.


B.

Because Type 1 changes overwrite existing records, Delta Lake cannot guarantee data consistency for cloned tables.


C.

The metadata created by the clone operation is referencing data files that were purged as invalid by the vacuum command


D.

Running vacuum automatically invalidates any shallow clones of a table; deep clone should always be used when a cloned table will be repeatedly queried.


Get Premium Databricks-Certified-Professional-Data-Engineer Questions

Contribute your Thoughts:


Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.