ExamGecko
Home Home / Snowflake / DEA-C01

Snowflake DEA-C01 Practice Test - Questions Answers, Page 9

Question list
Search
Search

List of questions

Search

Data engineer designed the data pipelines using Snowpipe to load data files into Snowflake tables, what will happen in case few files with same name but modified data are queued for reloading?

A.
Data will be reloaded as files are modified & its associated metadata also changed. But Snowflake handle implicitly deduplication.
A.
Data will be reloaded as files are modified & its associated metadata also changed. But Snowflake handle implicitly deduplication.
Answers
B.
eTAG is changed for Files even they are having same name, so data will be duplicated in SnowFlake tables.
B.
eTAG is changed for Files even they are having same name, so data will be duplicated in SnowFlake tables.
Answers
C.
Snowpipe uses file loading metadata associated with each table object, so no metadata available to prevent duplication.
C.
Snowpipe uses file loading metadata associated with each table object, so no metadata available to prevent duplication.
Answers
D.
Snowpipe uses file loading metadata associated with each pipe object to prevent reload-ing the same files (and duplicating data) in a table.
D.
Snowpipe uses file loading metadata associated with each pipe object to prevent reload-ing the same files (and duplicating data) in a table.
Answers
Suggested answer: D

Explanation:

Snowflake uses file loading metadata to prevent reloading the same files (and duplicating data) in a table. Snowpipe prevents loading files with the same name even if they were later modified (i.e.

have a different eTag).

The file loading metadata is associated with the pipe object rather than the table. As a result:

· Staged files with the same name as files that were already loaded are ignored, even if they have been modified, e.g. if new rows were added or errors in the file were corrected.

· Truncating the table using the TRUNCATE TABLE command does not delete the Snowpipe file loading metadata.

Data Engineer decided to call the public REST endpoints to load data and retrieve load history reports.

Which of the following REST endpoints and a Snowflake Information Schema table function for viewing your load history can be used by her? [Select All that apply]

A.
REST endpoints: loadinsertReport or HistoryScan
A.
REST endpoints: loadinsertReport or HistoryScan
Answers
B.
Information Schema table function: COPY_HISTORY
B.
Information Schema table function: COPY_HISTORY
Answers
C.
Account Usage view:COPY_HISTORY
C.
Account Usage view:COPY_HISTORY
Answers
D.
REST endpoints: insertReport or loadHistoryScan
D.
REST endpoints: insertReport or loadHistoryScan
Answers
E.
Information Schema table function: LOAD_HISTORY
E.
Information Schema table function: LOAD_HISTORY
Answers
Suggested answer: B, C, D

Explanation:

Snowflake provides REST endpoints and an Snowflake Information Schema table function for viewing your load history:

REST endpoints:

insertReport

loadHistoryScan

Information Schema table function:

COPY_HISTORY

Account Usage view:

COPY_HISTORY

Mark the incorrect statement when Data Engineer implement Automating Continuous Data Loading Using Cloud Messaging?

A.
Automated Snowpipe uses event notifications to determine when new files arrive in monitored cloud storage and are ready to load.
A.
Automated Snowpipe uses event notifications to determine when new files arrive in monitored cloud storage and are ready to load.
Answers
B.
When a pipe is paused, event messages received for the pipe enter a limited retention period. The period is 14 days by default. If a pipe is paused for longer than 14 days, it is considered stale.
B.
When a pipe is paused, event messages received for the pipe enter a limited retention period. The period is 14 days by default. If a pipe is paused for longer than 14 days, it is considered stale.
Answers
C.
Notifications identify the cloud storage event and include a list of the file names. They do not include the actual data in the files.
C.
Notifications identify the cloud storage event and include a list of the file names. They do not include the actual data in the files.
Answers
D.
Triggering automated Snowpipe data loads using S3 event messages is supported by Snowflake accounts hosted on Cloud Platform like AWS, GCP or AZURE.
D.
Triggering automated Snowpipe data loads using S3 event messages is supported by Snowflake accounts hosted on Cloud Platform like AWS, GCP or AZURE.
Answers
Suggested answer: D

Explanation:

Triggering automated Snowpipe data loads using S3 event messages is supported by Snowflake accounts hosted on Amazon Web Services (AWS) only.

Rest is correct statements.

Snowpipe API provides a REST endpoint for defining the list of files to ingest that Informs Snow-flake about the files to be ingested into a table. A successful response from this endpoint means that

Snowflake has recorded the list of files to add to the table. It does not necessarily mean the files have been ingested. What is name of this Endpoint?

A.
REST endpoints --> insertReport
A.
REST endpoints --> insertReport
Answers
B.
REST endpoints --> loadHistoryScan
B.
REST endpoints --> loadHistoryScan
Answers
C.
REST endpoints --> ingestfiles
C.
REST endpoints --> ingestfiles
Answers
D.
REST endpoints--> insertfiles
D.
REST endpoints--> insertfiles
Answers
Suggested answer: D

Explanation:

The Snowpipe API provides a REST endpoint for defining the list of files to ingest.

Endpoint: insertFiles

Informs Snowflake about the files to be ingested into a table. A successful response from this endpoint means that Snowflake has recorded the list of files to add to the table. It does not necessarily mean the files have been ingested. For more details, see the response codes below.

In most cases, Snowflake inserts fresh data into the target table within a few minutes.

To Know more about SnowFlake Rest API used for Data File ingestion, do refer:

https://docs.snowflake.com/en/user-guide/data-load-snowpipe-rest-apis.html#data-file-ingestion

The Snowpipe API provides REST endpoints for fetching load reports. One of the Endpoint named insertReport helps to retrieves a report of files submitted via insertFiles end point whose contents were recently ingested into a table. A success response (200) contains information about files that have recently been added to the table. Response Looks like below:

A.
{
A.
{
Answers
B.
"pipe": "SNOWTESTDB.SFTESTSCHEMA.SFpipe",
B.
"pipe": "SNOWTESTDB.SFTESTSCHEMA.SFpipe",
Answers
C.
"completeResult": true,
C.
"completeResult": true,
Answers
D.
"nextBeginMark": "1_16",
D.
"nextBeginMark": "1_16",
Answers
E.
"files": [
E.
"files": [
Answers
F.
{
F.
{
Answers
G.
"path": "data4859992083898.csv",
G.
"path": "data4859992083898.csv",
Answers
H.
"stageLocation": "s3://mybucket/",
H.
"stageLocation": "s3://mybucket/",
Answers
I.
"fileSize": 89,10. "timeReceived": "2022-01-31T04:47:41.453Z",11. "lastInsertTime": "2022-01-31T04:48:28.575Z",12. "rowsInserted": 1,13. "rowsParsed": 1,14. "errorsSeen": 0,15. "errorLimit": 1,16. "complete": true,17. "status": "????"18. }19. ]20. }Which one is the correct value of status string data in the Response Body?
I.
"fileSize": 89,10. "timeReceived": "2022-01-31T04:47:41.453Z",11. "lastInsertTime": "2022-01-31T04:48:28.575Z",12. "rowsInserted": 1,13. "rowsParsed": 1,14. "errorsSeen": 0,15. "errorLimit": 1,16. "complete": true,17. "status": "????"18. }19. ]20. }Which one is the correct value of status string data in the Response Body?
Answers
J.
LOADED
J.
LOADED
Answers
K.
LOADED_SUCCESS
K.
LOADED_SUCCESS
Answers
L.
LOAD_SUCCESS
L.
LOAD_SUCCESS
Answers
M.
SUCCESS
M.
SUCCESS
Answers
Suggested answer: C

Explanation:

Permissible Load status for the file:

LOAD_IN_PROGRESS: Part of the file has been loaded into the table, but the load process has not completed yet.

LOADED: The entire file has been loaded successfully into the table.

LOAD_FAILED: The file load failed.

PARTIALLY_LOADED: Some rows from this file were loaded successfully, but others were not loaded due to errors. Processing of this file is completed.

Please not the different Response Codes available with their meaning.

200 — Success. Report returned.

400 — Failure. Invalid request due to an invalid format, or limit exceeded.

404 — Failure. pipeName not recognized.

This error code can also be returned if the role used when calling the endpoint does not have sufficient privileges. For more information, see Granting Access Privileges.

429 — Failure. Request rate limit exceeded.

500 — Failure. Internal error occurred.

As you could understand from the questions, there is 200 Success response returned, Status in the response body would be LOADED.

Data Engineer try to load data from external stage using Snowpipe & later find out that some Set of Files Not Loaded. To debug the issue, she used COPY_HISTORY function & cross verified that its output indicates a subset of files was not loaded. What is possible reason of arising this situation in both REST API call and Auto-Ingest methods? [Select 2]

A.
External event-driven functionality is used to call the REST APIs, and a backlog of da-ta files already existed in the external stage before the events were configured.
A.
External event-driven functionality is used to call the REST APIs, and a backlog of da-ta files already existed in the external stage before the events were configured.
Answers
B.
An event notification failure prevented a set of files from getting queued.
B.
An event notification failure prevented a set of files from getting queued.
Answers
C.
Files modified and staged again after 14 days and Snowpipe ignores modified files that are staged again.
C.
Files modified and staged again after 14 days and Snowpipe ignores modified files that are staged again.
Answers
D.
A backlog of data files already existed in the external stage do not have any impact on Load failure, as this is well managed by serverless SnowPipe
D.
A backlog of data files already existed in the external stage do not have any impact on Load failure, as this is well managed by serverless SnowPipe
Answers
Suggested answer: A, B

Explanation:

COPY_HISTORY Record Indicates Unloaded Subset of Files:

If the COPY_HISTORY function output indicates a subset of files was not loaded, you may try to "refresh" the pipe.

This situation can arise in any of the following situations:

· The external stage was previously used to bulk load data using the COPY INTO table command.

· REST API:

o External event-driven functionality is used to call the REST APIs, and a backlog of data files al-ready existed in the external stage before the events were configured.

· Auto-ingest:

o A backlog of data files already existed in the external stage before event notifications were configured.

o An event notification failure prevented a set of files from getting queued.

To load the data files in your external stage using the configured pipe, execute an ALTER PIPE … REFRESH statement.

Data Engineer is looking out to delete staged files automatically/periodically when the data is successfully loaded into tables by the Snowpipe. For achieving the same, which options/command is best suited: [Select 2]

A.
PURGE option can be set as True in the COPY INTO Command embedded in PIPE objects definition.
A.
PURGE option can be set as True in the COPY INTO Command embedded in PIPE objects definition.
Answers
B.
To remove staged files that no longer needed, periodically REMOVE command can be executed to delete the files.
B.
To remove staged files that no longer needed, periodically REMOVE command can be executed to delete the files.
Answers
C.
To remove staged files that no longer needed, periodically DELETE command can be executed to delete the files.
C.
To remove staged files that no longer needed, periodically DELETE command can be executed to delete the files.
Answers
D.
REMOVE_STAGE_FILES option can be set as True in the COPY INTO Command embedded in PIPE objects definition.
D.
REMOVE_STAGE_FILES option can be set as True in the COPY INTO Command embedded in PIPE objects definition.
Answers
Suggested answer: A, B

Explanation:

Deleting Staged Files After Snowpipe Loads the Data

Pipe objects do not support the PURGE copy option. Snowpipe cannot delete staged files automatically when the data is successfully loaded into tables.

To remove staged files that you no longer need, It is recommended to periodically executing the REMOVE command to delete the files.

Alternatively, configure any lifecycle management features provided by cloud storage service provider.

Data Engineer is using existing pipe that automates data loads using event notifications, later he figured out the needs to modify pipe properties. For the same, He decided to recreate the pipe as best practice. He followed the below steps for the same.

A.
Query the SYSTEM$PIPE_STATUS function and verify that the pipe execution state is RUN-NING.
A.
Query the SYSTEM$PIPE_STATUS function and verify that the pipe execution state is RUN-NING.
Answers
B.
Recreate the pipe (using CREATE OR REPLACE PIPE).
B.
Recreate the pipe (using CREATE OR REPLACE PIPE).
Answers
C.
Query the SYSTEM$PIPE_STATUS function and verify that the pipe execution state is RUN-NING.Which are the Missing recommended steps while Recreating Pipes for Automated Data Loads?
C.
Query the SYSTEM$PIPE_STATUS function and verify that the pipe execution state is RUN-NING.Which are the Missing recommended steps while Recreating Pipes for Automated Data Loads?
Answers
D.
CREATE OR REPLACE PIPE command will recreate the PIPE successfully.
D.
CREATE OR REPLACE PIPE command will recreate the PIPE successfully.
Answers
E.
Terminate the existing pipe (using ALTER PIPE … SET PIPE_EXECUTION_TERMINATE = true) before recreation.
E.
Terminate the existing pipe (using ALTER PIPE … SET PIPE_EXECUTION_TERMINATE = true) before recreation.
Answers
F.
Pause the pipe (using ALTER PIPE … SET PIPE_EXECUTION_PAUSED = true) Pre & Post recreation & Resume after recreation (using ALTER PIPE … SET PIPE_EXECUTION_PAUSED = false).
F.
Pause the pipe (using ALTER PIPE … SET PIPE_EXECUTION_PAUSED = true) Pre & Post recreation & Resume after recreation (using ALTER PIPE … SET PIPE_EXECUTION_PAUSED = false).
Answers
G.
Force the pipe to resume (using SYSTEM$PIPE_FORCE_RESUME).
G.
Force the pipe to resume (using SYSTEM$PIPE_FORCE_RESUME).
Answers
Suggested answer: C

Explanation:

Recreating a pipe (using a CREATE OR REPLACE PIPE statement) is necessary to modify most pipe properties.

Recreating Pipes for Automated Data Loads When recreating a pipe that automates data loads using event notifications, it's recommended that Data Engineer complete the following steps:

1. Pause the pipe (using ALTER PIPE … SET PIPE_EXECUTION_PAUSED = true).

2. Query the SYSTEM$PIPE_STATUS function and verify that the pipe execution state is PAUSED.

3. Recreate the pipe (using CREATE OR REPLACE PIPE).

4. Pause the pipe again.

5. Review the configuration steps for your cloud messaging service to ensure the settings are still accurate.

6. Query the SYSTEM$PIPE_STATUS function again and verify that the pipe execution state is RUNNING.

Dominic, a Data Engineer wants to resume the pipe named stalepipe3 which got stale after 14 days.

To do the same, he called the SYSTEM$PIPE_FORCE_RESUME function select system$ pipe_force_resume('snowmydb.mysnowschema.stalepipe3','staleness_check_override');

Let's say If the pipe is resumed 16 days after it was paused, what will happened to the event notification that were received on the first and second days after the pipe was paused?

A.
Snowpipe generally skips any event notifications that were received on the first and second days after the pipe was paused.
A.
Snowpipe generally skips any event notifications that were received on the first and second days after the pipe was paused.
Answers
B.
Pipe maintains Metadata history of files for 64 days, so in this scenarios Snowpipe pro-cessed all the event notifications that were received for 16 days or so.
B.
Pipe maintains Metadata history of files for 64 days, so in this scenarios Snowpipe pro-cessed all the event notifications that were received for 16 days or so.
Answers
C.
Once the Pipe got stale, all the events got purged automatically & pipe needs to be rec-reated with modified properties.
C.
Once the Pipe got stale, all the events got purged automatically & pipe needs to be rec-reated with modified properties.
Answers
D.
All the events get processed from day 1 if the PURGE properties in the PIPE object definition set to be FALSE initially.
D.
All the events get processed from day 1 if the PURGE properties in the PIPE object definition set to be FALSE initially.
Answers
Suggested answer: A

Explanation:

When a pipe is paused, event messages received for the pipe enter a limited retention period. The period is 14 days by default. If a pipe is paused for longer than 14 days, it is considered stale.

To resume a stale pipe, a qualified role must call the SYSTEM$PIPE_FORCE_RESUME function and input the STALENESS_CHECK_OVERRIDE argument. This argument indicates an under-standing that the role is resuming a stale pipe.

For example, resume the stale stalepipe1 pipe in the mydb.myschema database and schema:

select sys-tem$pipe_force_resume('mydb.myschema.stalepipe3','staleness_check_override');

As an event notification received while a pipe is paused reaches the end of the limited retention period, Snowflake schedules it to be dropped from the internal metadata. If the pipe is later resumed,

Snowpipe processes these older notifications on a best effort basis. Snowflake cannot guarantee that they are processed.

For example, if a pipe is resumed 15 days after it was paused, Snowpipe generally skips any event notifications that were received on the first day the pipe was paused (i.e. that are now more than 14 days old). If the pipe is resumed 16 days after it was paused, Snowpipe generally skips any event notifications that were received on the first and second days after the pipe was paused. And so on.

Data Engineer looking out for quick tool for understanding the mechanics of queries & need to know more about the performance or behaviour of a particular query.

He should go to which feature of snowflake which can help him to spot typical mistakes in SQL query expressions to identify potential performance bottlenecks and improvement opportunities?

A.
Query Optimizer
A.
Query Optimizer
Answers
B.
Performance Metadata table
B.
Performance Metadata table
Answers
C.
Query Profile
C.
Query Profile
Answers
D.
Query Designer
D.
Query Designer
Answers
Suggested answer: C

Explanation:

Query Profile, available through the classic web interface, provides execution details for a query. For the selected query, it provides a graphical representation of the main components of the pro-cessing plan for the query, with statistics for each component, along with details and statistics for the overall query.

Query Profile is a powerful tool for understanding the mechanics of queries. It can be used whenever you want or need to know more about the performance or behavior of a particular query. It is designed to help you spot typical mistakes in SQL query expressions to identify potential performance bottlenecks and improvement opportunities.

Total 130 questions
Go to page: of 13