ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 5 - Professional Data Engineer discussion

Report
Export

An external customer provides you with a daily dump of data from their database. The data flows into Google Cloud Storage GCS as comma-separated values (CSV) files. You want to analyze this data in Google BigQuery, but the data could have rows that are formatted incorrectly or corrupted. How should you build this pipeline?

A.
Use federated data sources, and check data in the SQL query.
Answers
A.
Use federated data sources, and check data in the SQL query.
B.
Enable BigQuery monitoring in Google Stackdriver and create an alert.
Answers
B.
Enable BigQuery monitoring in Google Stackdriver and create an alert.
C.
Import the data into BigQuery using the gcloud CLI and set max_bad_records to 0.
Answers
C.
Import the data into BigQuery using the gcloud CLI and set max_bad_records to 0.
D.
Run a Google Cloud Dataflow batch pipeline to import the data into BigQuery, and push errors to another dead-letter table for analysis.
Answers
D.
Run a Google Cloud Dataflow batch pipeline to import the data into BigQuery, and push errors to another dead-letter table for analysis.
Suggested answer: D
asked 18/09/2024
Neville Raposo
37 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first