Snowflake SnowPro Advanced: Data Engineer Certification Exam DEA-C01 Question # 8 Topic 1 Discussion

Snowflake SnowPro Advanced: Data Engineer Certification Exam DEA-C01 Question # 8 Topic 1 Discussion

DEA-C01 Exam Topic 1 Question 8 Discussion:
Question #: 8
Topic #: 1

A CSV file around 1 TB in size is generated daily on an on-premise server A corresponding table. Internal stage, and file format have already been created in Snowflake to facilitate the data loading process

How can the process of bringing the CSV file into Snowflake be automated using the LEAST amount of operational overhead?


A.

Create a task in Snowflake that executes once a day and runs a copy into statement that references the internal stage The internal stage will read the files directly

from the on-premise server and copy the newest file into the table from the on-premise server to the Snowflake table


B.

On the on-premise server schedule a SQL file to run using SnowSQL that executes a PUT to push a specific file to the internal stage Create a task that executes once a

day m Snowflake and runs a OOPY WTO statement that references the internal stage Schedule the task to start after the file lands in the internal stage


C.

On the on-premise server schedule a SQL file to run using SnowSQL that executes a PUT to push a specific file to the internal stage. Create a pipe that runs a copy

into statement that references the internal stage Snowpipe auto-ingest will automatically load the file from the internal stage when the new file lands in the internal

stage.


D.

On the on premise server schedule a Python file that uses the Snowpark Python library. The Python script will read the CSV data into a DataFrame and generate an

insert into statement that will directly load into the table The script will bypass the need to move a file into an internal stage


Get Premium DEA-C01 Questions

Contribute your Thoughts:


Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.