New Year Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 70percent

Snowflake DEA-C01 SnowPro Advanced: Data Engineer Certification Exam Exam Practice Test

Demo: 17 questions
Total 65 questions

SnowPro Advanced: Data Engineer Certification Exam Questions and Answers

Question 1

A Data Engineer enables a result cache at the session level with the following command:

ALTER SESSION SET USE CACHED RESULT = TRUE;

The Engineer then runs the following select query twice without delay:

The underlying table does not change between executions

What are the results of both runs?

Options:

A.

The first and second run returned the same results because sample is deterministic

B.

The first and second run returned the same results, because the specific SEEDvalue was provided.

C.

The first and second run returned different results because the query is evaluated each time it is run.

D.

The first and second run returned differentresults because the query uses *instead of an explicit column list

Question 2

A large table with 200 columns contains two years of historical data. When queried. the table is filtered on a single day Below is the Query Profile:

Using a size 2XL virtual warehouse, this query look over an hour to complete

What will improve the query performance the MOST?

Options:

A.

increase the size of the virtual warehouse.

B.

Increase the number of clusters in the virtual warehouse

C.

Implement the search optimization service on the table

D.

Add a date column as a cluster key on the table

Question 3

A new customer table is created by a data pipeline in a Snowflake schema where MANAGED ACCESSenabled.

…. Can gran access to the CUSTOMER table? (Select THREE.)

Options:

A.

The role that owns the schema

B.

The role that owns the database

C.

The role that owns the customer table

D.

The SYSADMIN role

E.

The SECURITYADMIN role

F.

The USERADMIN role with the manage grants privilege

Question 4

Which methods will trigger an action that will evaluate a DataFrame? (Select TWO)

Options:

A.

DataFrame.random_split ( )

B.

DataFrame.collect ()

C.

DateFrame.select ()

D.

DataFrame.col ( )

E.

DataFrame.show ()

Question 5

Which system role is recommended for a custom role hierarchy to be ultimately assigned to?

Options:

A.

ACCOUNTADMIN

B.

SECURITYADMIN

C.

SYSTEMADMIN

D.

USERADMIN

Question 6

When would a Data engineer use table with the flatten function instead of the lateral flatten combination?

Options:

A.

When TABLE with FLATTENrequires another source in the from clause to refer to

B.

WhenTABLE with FLATTENrequires no additional source m the from clause to refer to

C.

Whenthe LATERALFLATTENcombination requires no other source m the from clause to refer to

D.

When table withFLATTENis acting like a sub-query executed for each returned row

Question 7

A stream called TRANSACTIONS_STM is created on top of a transactions table in a continuous pipeline running in Snowflake. After a couple of months, the TRANSACTIONS table is renamed transactiok3_raw to comply with new naming standards

What will happen to the TRANSACTIONS _STM object?

Options:

A.

TRANSACTIONS _STMwill keep working as expected

B.

TRANSACTIONS _STMwill be stale and will need to be re-created

C.

TRANSACTIONS _STMwill be automatically renamedTRANSACTIONS _RAW_STM.

D.

Reading from the traksactioks_3T>: stream will succeed for some time after the expected STALE_TIME.

Question 8

A CSV file around 1 TB in size is generated daily on an on-premise server A corresponding table. Internal stage, and file format have already been created in Snowflake to facilitate the data loading process

How can the process of bringing the CSV file into Snowflake be automated using the LEAST amount of operational overhead?

Options:

A.

Create a task in Snowflake that executes once a day and runs a copy into statement that references the internal stage The internal stage will read the files directly

from the on-premise server and copy the newest file into the table from the on-premise server to the Snowflake table

B.

On the on-premise server schedule a SQL file to run using SnowSQL that executes a PUT to push a specific file to the internal stage Create a task that executes once a

day m Snowflake and runs a OOPY WTO statement that references the internal stage Schedule the task to start after the file lands in the internal stage

C.

On the on-premise server schedule a SQL file to run using SnowSQL that executes a PUT to push a specific file to the internal stage. Create a pipe that runs a copy

into statement that references the internal stage Snowpipe auto-ingest will automatically load the file from the internal stage when the new file lands in the internal

stage.

D.

On the on premise server schedule a Python file that uses the Snowpark Python library. The Python script will read the CSV data into a DataFrame and generate an

insert into statement that will directly load into the table The script will bypass the need to move a file into an internal stage

Question 9

A company is building a dashboard for thousands of Analysts. The dashboard presents the results of a few summary queries on tables that are regularly updated. The query conditions vary by tope according to what data each Analyst needs Responsiveness of the dashboard queries is a top priority, and the data cache should be preserved.

How should the Data Engineer configure the compute resources to support this dashboard?

Options:

A.

Assign queries to a multi-cluster virtual warehouse with economy auto-scaling Allow the system to automatically start and stop clusters according to demand.

B.

Assign all queries to a multi-cluster virtual warehouse set to maximized mode Monitor to determine the smallest suitable number of clusters.

C.

Create a virtual warehouse for every 250 Analysts Monitor to determine how many of these virtual warehouses are being utilized at capacity.

D.

Create a size XL virtual warehouse to support all the dashboard queries Monitor query runtimes to determine whether the virtual warehouse should be resized.

Question 10

Which callback function is required within a JavaScript User-Defined Function (UDF) for it to execute successfully?

Options:

A.

initialize ()

B.

processRow ()

C.

handler

D.

finalize ()

Question 11

A Data Engineer is writing a Python script using the Snowflake Connector for Python. The Engineer will use the snowflake. Connector.connect function to connect to Snowflake The requirementsare:

*Raise an exception if the specified database schema or warehouse does not exist

*improve download performance

Whichparameters of the connect function should be used? (Select TWO).

Options:

A.

authenticator

B.

arrow_nunber_to_decimal

C.

client_prefetch_threads

D.

client_session_keep_alivs

E.

validate_default_parameters

Question 12

Which output is provided by both theSYSTEM$CLUSTERING_DEPTHfunction and theSYSTEM$CLUSTERING_INFORMATIONfunction?

Options:

A.

average_depth

B.

notes

C.

average_overlaps

D.

total_partition_count

Question 13

Assuming that the session parameter USE_CACHED_RESULT is set to false, what are characteristics of Snowflake virtual warehouses in terms of the use of Snowpark?

Options:

A.

Creating a DataFrame from a table will start a virtual warehouse

B.

Creating a DataFrame from a staged file with the read () method will start a virtual warehouse

C.

Transforming a DataFrame with methods like replace () will start a virtual warehouse -

D.

Calling a Snowpark stored procedure to query the database with session, call () will start a virtual warehouse

Question 14

Which methods can be used to create a DataFrame object in Snowpark? (Select THREE)

Options:

A.

session.jdbc_connection()

B.

session.read.json{)

C.

session,table()

D.

DataFraas.writeO

E.

session.builder()

F.

session.sql()

Question 15

A company is using Snowpipe to bring in millions of rows every day of Change Data Capture (CDC) into a Snowflake staging table on a real-time basis The CDC needs to get processedand combined with other data in Snowflake and land in a final table as part of the full data pipeline.

How can a Data engineer MOST efficiently process the incoming CDC on an ongoing basis?

Options:

A.

Create a stream on the staging table and schedule a task that transforms data from the stream only when the stream has data.

B.

Transform the data during the data load with Snowpipe by modifying the related copy into statement to include transformation steps such as case statements andJOIN'S.

C.

Schedule a task that dynamically retrieves the last time the task was run from information_schema-rask_hiSwOry and use that timestamp to process the delta of the new rows since the last time the task was run.

D.

Use a create ok replace table as statement that references the staging table and includes all the transformation SQL. Use a task to run the full create or replace table as statement on a scheduled basis

Question 16

A Data Engineer is investigating a query that is taking a long time to return The Query Profile shows the following:

What step should the Engineer take to increase the query performance?

Options:

A.

Add additional virtual warehouses.

B.

increasethe size of the virtual warehouse.

C.

Rewrite the query using Common Table Expressions (CTEs)

D.

Change the order of the joins and start with smaller tables first

Question 17

A secure function returns data coming through an inbound share

What will happen if a Data Engineer tries to assign usage privileges on this function to an outbound share?

Options:

A.

An error will be returned because the Engineer cannot share data that has already been shared

B.

An error will be returned because only views and secure stored procedures can be shared

C.

An error will be returned because only secure functions can be shared with inbound shares

D.

The Engineer will be able to share the secure function with other accounts

Demo: 17 questions
Total 65 questions