Pass the Snowflake SnowPro Advanced: Architect ARA-C01 Questions and answers with CertsForce

Viewing page 2 out of 5 pages
Viewing questions 11-20 out of questions
Questions # 11:

A company's Architect needs to find an efficient way to get data from an external partner, who is also a Snowflake user. The current solution is based on daily JSON extracts that are placed on an FTP server and uploaded to Snowflake manually. The files are changed several times each month, and the ingestion process needs to be adapted to accommodate these changes.

What would be the MOST efficient solution?

Options:

A.

Ask the partner to create a share and add the company's account.


B.

Ask the partner to use the data lake export feature and place the data into cloud storage where Snowflake can natively ingest it (schema-on-read).


C.

Keep the current structure but request that the partner stop changing files, instead only appending new files.


D.

Ask the partner to set up a Snowflake reader account and use that account to get the data for ingestion.


Expert Solution
Questions # 12:

Which columns can be included in an external table schema? (Select THREE).

Options:

A.

VALUE


B.

METADATASROW_ID


C.

METADATASISUPDATE


D.

METADAT A$ FILENAME


E.

METADATAS FILE_ROW_NUMBER


F.

METADATASEXTERNAL TABLE PARTITION


Expert Solution
Questions # 13:

What Snowflake features should be leveraged when modeling using Data Vault?

Options:

A.

Snowflake’s support of multi-table inserts into the data model’s Data Vault tables


B.

Data needs to be pre-partitioned to obtain a superior data access performance


C.

Scaling up the virtual warehouses will support parallel processing of new source loads


D.

Snowflake’s ability to hash keys so that hash key joins can run faster than integer joins


Expert Solution
Questions # 14:

A media company needs a data pipeline that will ingest customer review data into a Snowflake table, and apply some transformations. The company also needs to use Amazon Comprehend to do sentiment analysis and make the de-identified final data set available publicly for advertising companies who use different cloud providers in different regions.

The data pipeline needs to run continuously ang efficiently as new records arrive in the object storage leveraging event notifications. Also, the operational complexity, maintenance of the infrastructure, including platform upgrades and security, and the development effort should be minimal.

Which design will meet these requirements?

Options:

A.

Ingest the data using COPY INTO and use streams and tasks to orchestrate transformations. Export the data into Amazon S3 to do model inference with Amazon Comprehend and ingest the data back into a Snowflake table. Then create a listing in the Snowflake Marketplace to make the data available to other companies.


B.

Ingest the data using Snowpipe and use streams and tasks to orchestrate transformations. Create an external function to do model inference with Amazon Comprehend and write the final records to a Snowflake table. Then create a listing in the Snowflake Marketplace to make the data available to other companies.


C.

Ingest the data into Snowflake using Amazon EMR and PySpark using the Snowflake Spark connector. Apply transformations using another Spark job. Develop a python program to do model inference by leveraging the Amazon Comprehend text analysis API. Then write the results to a Snowflake table and create a listing in the Snowflake Marketplace to make the data available to other companies.


D.

Ingest the data using Snowpipe and use streams and tasks to orchestrate transformations. Export the data into Amazon S3 to do model inference with Amazon Comprehend and ingest the data back into a Snowflake table. Then create a listing in the Snowflake Marketplace to make the data available to other companies.


Expert Solution
Questions # 15:

Which Snowflake objects can be used in a data share? (Select TWO).

Options:

A.

Standard view


B.

Secure view


C.

Stored procedure


D.

External table


E.

Stream


Expert Solution
Questions # 16:

An Architect is designing a pipeline to stream event data into Snowflake using the Snowflake Kafka connector. The Architect’s highest priority is to configure the connector to stream data in the MOST cost-effective manner.

Which of the following is recommended for optimizing the cost associated with the Snowflake Kafka connector?

Options:

A.

Utilize a higher Buffer.flush.time in the connector configuration.


B.

Utilize a higher Buffer.size.bytes in the connector configuration.


C.

Utilize a lower Buffer.size.bytes in the connector configuration.


D.

Utilize a lower Buffer.count.records in the connector configuration.


Expert Solution
Questions # 17:

What is a characteristic of loading data into Snowflake using the Snowflake Connector for Kafka?

Options:

A.

The Connector only works in Snowflake regions that use AWS infrastructure.


B.

The Connector works with all file formats, including text, JSON, Avro, Ore, Parquet, and XML.


C.

The Connector creates and manages its own stage, file format, and pipe objects.


D.

Loads using the Connector will have lower latency than Snowpipe and will ingest data in real time.


Expert Solution
Questions # 18:

What is a characteristic of event notifications in Snowpipe?

Options:

A.

The load history is stored In the metadata of the target table.


B.

Notifications identify the cloud storage event and the actual data in the files.


C.

Snowflake can process all older notifications when a paused pipe Is resumed.


D.

When a pipe Is paused, event messages received for the pipe enter a limited retention period.


Expert Solution
Questions # 19:

A company wants to Integrate its main enterprise identity provider with federated authentication with Snowflake.

The authentication integration has been configured and roles have been created in Snowflake. However, the users are not automatically appearing in Snowflake when created and their group membership is not reflected in their assigned rotes.

How can the missing functionality be enabled with the LEAST amount of operational overhead?

Options:

A.

OAuth must be configured between the identity provider and Snowflake. Then the authorization server must be configured with the right mapping of users and roles.


B.

OAuth must be configured between the identity provider and Snowflake. Then the authorization server must be configured with the right mapping of users, and the resource server must be configured with the right mapping of role assignment.


C.

SCIM must be enabled between the identity provider and Snowflake. Once both are synchronized through SCIM, their groups will get created as group accounts in Snowflake and the proper roles can be granted.


D.

SCIM must be enabled between the identity provider and Snowflake. Once both are synchronized through SCIM. users will automatically get created and their group membership will be reflected as roles In Snowflake.


Expert Solution
Questions # 20:

An Architect Is designing a data lake with Snowflake. The company has structured, semi-structured, and unstructured data. The company wants to save the data inside the data lake within the Snowflake system. The company is planning on sharing data among Its corporate branches using Snowflake data sharing.

What should be considered when sharing the unstructured data within Snowflake?

Options:

A.

A pre-signed URL should be used to save the unstructured data into Snowflake in order to share data over secure views, with no time limit for the URL.


B.

A scoped URL should be used to save the unstructured data into Snowflake in order to share data over secure views, with a 24-hour time limit for the URL.


C.

A file URL should be used to save the unstructured data into Snowflake in order to share data over secure views, with a 7-day time limit for the URL.


D.

A file URL should be used to save the unstructured data into Snowflake in order to share data over secure views, with the "expiration_time" argument defined for the URL time limit.


Expert Solution
Viewing page 2 out of 5 pages
Viewing questions 11-20 out of questions