ExamGecko
Home / Google / Professional Data Engineer / List of questions
Ask Question

Google Professional Data Engineer Practice Test - Questions Answers, Page 11

List of questions

Question 101

Report
Export
Collapse

What are the minimum permissions needed for a service account used with Google Dataproc?

Execute to Google Cloud Storage; write to Google Cloud Logging
Execute to Google Cloud Storage; write to Google Cloud Logging
Write to Google Cloud Storage; read to Google Cloud Logging
Write to Google Cloud Storage; read to Google Cloud Logging
Execute to Google Cloud Storage; execute to Google Cloud Logging
Execute to Google Cloud Storage; execute to Google Cloud Logging
Read and write to Google Cloud Storage; write to Google Cloud Logging
Read and write to Google Cloud Storage; write to Google Cloud Logging
Suggested answer: D

Explanation:

Service accounts authenticate applications running on your virtual machine instances to other Google Cloud Platform services. For example, if you write an application that reads and writes files on Google Cloud Storage, it must first authenticate to the Google Cloud Storage API. At a minimum, service accounts used with Cloud Dataproc need permissions to read and write to Google Cloud Storage, and to write to Google Cloud Logging.

Reference: https://cloud.google.com/dataproc/docs/concepts/service-accounts#important_notes

asked 18/09/2024
Antonio Agustin Mirano
34 questions

Question 102

Report
Export
Collapse

Which role must be assigned to a service account used by the virtual machines in a Dataproc cluster so they can execute jobs?

Dataproc Worker
Dataproc Worker
Dataproc Viewer
Dataproc Viewer
Dataproc Runner
Dataproc Runner
Dataproc Editor
Dataproc Editor
Suggested answer: A

Explanation:

Service accounts used with Cloud Dataproc must have Dataproc/Dataproc Worker role (or have all the permissions granted by Dataproc Worker role).

Reference: https://cloud.google.com/dataproc/docs/concepts/service-accounts#important_notes

asked 18/09/2024
Karen Vivanco
29 questions

Question 103

Report
Export
Collapse

When creating a new Cloud Dataproc cluster with the projects.regions.clusters.create operation, these four values are required: project, region, name, and ____.

zone
zone
node
node
label
label
type
type
Suggested answer: A

Explanation:

At a minimum, you must specify four values when creating a new cluster with the projects.regions.clusters.create operation:

The project in which the cluster will be created

The region to use

The name of the cluster

The zone in which the cluster will be created You can specify many more details beyond these minimum requirements. For example, you can also specify the number of workers, whether preemptible compute should be used, and the network settings.

Reference: https://cloud.google.com/dataproc/docs/tutorials/python-libraryexample#create_a_new_cloud_dataproc_cluste

asked 18/09/2024
Slavomir Ugrevic
36 questions

Question 104

Report
Export
Collapse

Which Google Cloud Platform service is an alternative to Hadoop with Hive?

Cloud Dataflow
Cloud Dataflow
Cloud Bigtable
Cloud Bigtable
BigQuery
BigQuery
Cloud Datastore
Cloud Datastore
Suggested answer: C

Explanation:

Apache Hive is a data warehouse software project built on top of Apache Hadoop for providing data summarization, query, and analysis.

Google BigQuery is an enterprise data warehouse.

Reference: https://en.wikipedia.org/wiki/Apache_Hive

asked 18/09/2024
haythem BRIGUI
31 questions

Question 105

Report
Export
Collapse

Which of these rules apply when you add preemptible workers to a Dataproc cluster (select 2 answers)?

Preemptible workers cannot use persistent disk.
Preemptible workers cannot use persistent disk.
Preemptible workers cannot store data.
Preemptible workers cannot store data.
If a preemptible worker is reclaimed, then a replacement worker must be added manually.
If a preemptible worker is reclaimed, then a replacement worker must be added manually.
A Dataproc cluster cannot have only preemptible workers.
A Dataproc cluster cannot have only preemptible workers.
Suggested answer: B, D

Explanation:

The following rules will apply when you use preemptible workers with a Cloud Dataproc cluster:

. Processing onlyóSince preemptibles can be reclaimed at any time, preemptible workers do not store data. Preemptibles added to a Cloud Dataproc cluster only function as processing nodes.

. No preemptible-only clustersóTo ensure clusters do not lose all workers, Cloud Dataproc cannot create preemptible-only clusters.

. Persistent disk sizeóAs a default, all preemptible workers are created with the smaller of 100GB or the primary worker boot disk size. This disk space is used for local caching of data and is not available through HDFS.

The managed group automatically re-adds workers lost due to reclamation as capacity permits.

Reference: https://cloud.google.com/dataproc/docs/concepts/preemptible-vms

asked 18/09/2024
Dimitri Bennett
35 questions

Question 106

Report
Export
Collapse

When using Cloud Dataproc clusters, you can access the YARN web interface by configuring a browser to connect through a ____ proxy.

HTTPS
HTTPS
VPN
VPN
SOCKS
SOCKS
HTTP
HTTP
Suggested answer: C

Explanation:

When using Cloud Dataproc clusters, configure your browser to use the SOCKS proxy. The SOCKS proxy routes data intended for the Cloud Dataproc cluster through an SSH tunnel.

Reference: https://cloud.google.com/dataproc/docs/concepts/cluster-web-interfaces#interfaces

asked 18/09/2024
Alex Rector
30 questions

Question 107

Report
Export
Collapse

Cloud Dataproc is a managed Apache Hadoop and Apache _____ service.

Blaze
Blaze
Spark
Spark
Fire
Fire
Ignite
Ignite
Suggested answer: B

Explanation:

Cloud Dataproc is a managed Apache Spark and Apache Hadoop service that lets you use open source data tools for batch processing, querying, streaming, and machine learning.

Reference: https://cloud.google.com/dataproc/docs/

asked 18/09/2024
Jeffrey Tiffany
43 questions

Question 108

Report
Export
Collapse

Which action can a Cloud Dataproc Viewer perform?

Submit a job.
Submit a job.
Create a cluster.
Create a cluster.
Delete a cluster.
Delete a cluster.
List the jobs.
List the jobs.
Suggested answer: D

Explanation:

A Cloud Dataproc Viewer is limited in its actions based on its role. A viewer can only list clusters, get cluster details, list jobs, get job details, list operations, and get operation details.

Reference:

https://cloud.google.com/dataproc/docs/concepts/iam#iam_roles_and_cloud_dataproc_operations_summary

asked 18/09/2024
Nathan Davies
38 questions

Question 109

Report
Export
Collapse

Dataproc clusters contain many configuration files. To update these files, you will need to use the -- properties option. The format for the option is: file_prefix:property=_____.

details
details
value
value
null
null
id
id
Suggested answer: B

Explanation:

To make updating files and properties easy, the --properties command uses a special format to specify the configuration file and the property and value within the file that should be updated. The formatting is as follows: file_prefix:property=value.

Reference: https://cloud.google.com/dataproc/docs/concepts/cluster-properties#formatting

asked 18/09/2024
Mpho Ntshontsi
40 questions

Question 110

Report
Export
Collapse

Scaling a Cloud Dataproc cluster typically involves ____.

increasing or decreasing the number of worker nodes
increasing or decreasing the number of worker nodes
increasing or decreasing the number of master nodes
increasing or decreasing the number of master nodes
moving memory to run more applications on a single node
moving memory to run more applications on a single node
deleting applications from unused nodes periodically
deleting applications from unused nodes periodically
Suggested answer: A

Explanation:

After creating a Cloud Dataproc cluster, you can scale the cluster by increasing or decreasing the number of worker nodes in the cluster at any time, even when jobs are running on the cluster. Cloud Dataproc clusters are typically scaled to:

1) increase the number of workers to make a job run faster

2) decrease the number of workers to save money

3) increase the number of nodes to expand available Hadoop Distributed Filesystem (HDFS) storage

Reference: https://cloud.google.com/dataproc/docs/concepts/scaling-clusters

asked 18/09/2024
Bianca Duizer
48 questions
Total 377 questions
Go to page: of 38

Related questions