ExamGecko
Home Home / Google / Professional Data Engineer

Google Professional Data Engineer Practice Test - Questions Answers, Page 11

Question list
Search
Search

Related questions











What are the minimum permissions needed for a service account used with Google Dataproc?

A.
Execute to Google Cloud Storage; write to Google Cloud Logging
A.
Execute to Google Cloud Storage; write to Google Cloud Logging
Answers
B.
Write to Google Cloud Storage; read to Google Cloud Logging
B.
Write to Google Cloud Storage; read to Google Cloud Logging
Answers
C.
Execute to Google Cloud Storage; execute to Google Cloud Logging
C.
Execute to Google Cloud Storage; execute to Google Cloud Logging
Answers
D.
Read and write to Google Cloud Storage; write to Google Cloud Logging
D.
Read and write to Google Cloud Storage; write to Google Cloud Logging
Answers
Suggested answer: D

Explanation:

Service accounts authenticate applications running on your virtual machine instances to other Google Cloud Platform services. For example, if you write an application that reads and writes files on Google Cloud Storage, it must first authenticate to the Google Cloud Storage API. At a minimum, service accounts used with Cloud Dataproc need permissions to read and write to Google Cloud Storage, and to write to Google Cloud Logging.

Reference: https://cloud.google.com/dataproc/docs/concepts/service-accounts#important_notes

Which role must be assigned to a service account used by the virtual machines in a Dataproc cluster so they can execute jobs?

A.
Dataproc Worker
A.
Dataproc Worker
Answers
B.
Dataproc Viewer
B.
Dataproc Viewer
Answers
C.
Dataproc Runner
C.
Dataproc Runner
Answers
D.
Dataproc Editor
D.
Dataproc Editor
Answers
Suggested answer: A

Explanation:

Service accounts used with Cloud Dataproc must have Dataproc/Dataproc Worker role (or have all the permissions granted by Dataproc Worker role).

Reference: https://cloud.google.com/dataproc/docs/concepts/service-accounts#important_notes

When creating a new Cloud Dataproc cluster with the projects.regions.clusters.create operation, these four values are required: project, region, name, and ____.

A.
zone
A.
zone
Answers
B.
node
B.
node
Answers
C.
label
C.
label
Answers
D.
type
D.
type
Answers
Suggested answer: A

Explanation:

At a minimum, you must specify four values when creating a new cluster with the projects.regions.clusters.create operation:

The project in which the cluster will be created

The region to use

The name of the cluster

The zone in which the cluster will be created You can specify many more details beyond these minimum requirements. For example, you can also specify the number of workers, whether preemptible compute should be used, and the network settings.

Reference: https://cloud.google.com/dataproc/docs/tutorials/python-libraryexample#create_a_new_cloud_dataproc_cluste

Which Google Cloud Platform service is an alternative to Hadoop with Hive?

A.
Cloud Dataflow
A.
Cloud Dataflow
Answers
B.
Cloud Bigtable
B.
Cloud Bigtable
Answers
C.
BigQuery
C.
BigQuery
Answers
D.
Cloud Datastore
D.
Cloud Datastore
Answers
Suggested answer: C

Explanation:

Apache Hive is a data warehouse software project built on top of Apache Hadoop for providing data summarization, query, and analysis.

Google BigQuery is an enterprise data warehouse.

Reference: https://en.wikipedia.org/wiki/Apache_Hive

Which of these rules apply when you add preemptible workers to a Dataproc cluster (select 2 answers)?

A.
Preemptible workers cannot use persistent disk.
A.
Preemptible workers cannot use persistent disk.
Answers
B.
Preemptible workers cannot store data.
B.
Preemptible workers cannot store data.
Answers
C.
If a preemptible worker is reclaimed, then a replacement worker must be added manually.
C.
If a preemptible worker is reclaimed, then a replacement worker must be added manually.
Answers
D.
A Dataproc cluster cannot have only preemptible workers.
D.
A Dataproc cluster cannot have only preemptible workers.
Answers
Suggested answer: B, D

Explanation:

The following rules will apply when you use preemptible workers with a Cloud Dataproc cluster:

. Processing onlyóSince preemptibles can be reclaimed at any time, preemptible workers do not store data. Preemptibles added to a Cloud Dataproc cluster only function as processing nodes.

. No preemptible-only clustersóTo ensure clusters do not lose all workers, Cloud Dataproc cannot create preemptible-only clusters.

. Persistent disk sizeóAs a default, all preemptible workers are created with the smaller of 100GB or the primary worker boot disk size. This disk space is used for local caching of data and is not available through HDFS.

The managed group automatically re-adds workers lost due to reclamation as capacity permits.

Reference: https://cloud.google.com/dataproc/docs/concepts/preemptible-vms

When using Cloud Dataproc clusters, you can access the YARN web interface by configuring a browser to connect through a ____ proxy.

A.
HTTPS
A.
HTTPS
Answers
B.
VPN
B.
VPN
Answers
C.
SOCKS
C.
SOCKS
Answers
D.
HTTP
D.
HTTP
Answers
Suggested answer: C

Explanation:

When using Cloud Dataproc clusters, configure your browser to use the SOCKS proxy. The SOCKS proxy routes data intended for the Cloud Dataproc cluster through an SSH tunnel.

Reference: https://cloud.google.com/dataproc/docs/concepts/cluster-web-interfaces#interfaces

Cloud Dataproc is a managed Apache Hadoop and Apache _____ service.

A.
Blaze
A.
Blaze
Answers
B.
Spark
B.
Spark
Answers
C.
Fire
C.
Fire
Answers
D.
Ignite
D.
Ignite
Answers
Suggested answer: B

Explanation:

Cloud Dataproc is a managed Apache Spark and Apache Hadoop service that lets you use open source data tools for batch processing, querying, streaming, and machine learning.

Reference: https://cloud.google.com/dataproc/docs/

Which action can a Cloud Dataproc Viewer perform?

A.
Submit a job.
A.
Submit a job.
Answers
B.
Create a cluster.
B.
Create a cluster.
Answers
C.
Delete a cluster.
C.
Delete a cluster.
Answers
D.
List the jobs.
D.
List the jobs.
Answers
Suggested answer: D

Explanation:

A Cloud Dataproc Viewer is limited in its actions based on its role. A viewer can only list clusters, get cluster details, list jobs, get job details, list operations, and get operation details.

Reference:

https://cloud.google.com/dataproc/docs/concepts/iam#iam_roles_and_cloud_dataproc_operations_summary

Dataproc clusters contain many configuration files. To update these files, you will need to use the -- properties option. The format for the option is: file_prefix:property=_____.

A.
details
A.
details
Answers
B.
value
B.
value
Answers
C.
null
C.
null
Answers
D.
id
D.
id
Answers
Suggested answer: B

Explanation:

To make updating files and properties easy, the --properties command uses a special format to specify the configuration file and the property and value within the file that should be updated. The formatting is as follows: file_prefix:property=value.

Reference: https://cloud.google.com/dataproc/docs/concepts/cluster-properties#formatting

Scaling a Cloud Dataproc cluster typically involves ____.

A.
increasing or decreasing the number of worker nodes
A.
increasing or decreasing the number of worker nodes
Answers
B.
increasing or decreasing the number of master nodes
B.
increasing or decreasing the number of master nodes
Answers
C.
moving memory to run more applications on a single node
C.
moving memory to run more applications on a single node
Answers
D.
deleting applications from unused nodes periodically
D.
deleting applications from unused nodes periodically
Answers
Suggested answer: A

Explanation:

After creating a Cloud Dataproc cluster, you can scale the cluster by increasing or decreasing the number of worker nodes in the cluster at any time, even when jobs are running on the cluster. Cloud Dataproc clusters are typically scaled to:

1) increase the number of workers to make a job run faster

2) decrease the number of workers to save money

3) increase the number of nodes to expand available Hadoop Distributed Filesystem (HDFS) storage

Reference: https://cloud.google.com/dataproc/docs/concepts/scaling-clusters

Total 372 questions
Go to page: of 38