New Year Special 75% Discount offer - Ends in 0d 00h 00m 00s - Coupon code: 75brite

Google Updated Professional-Data-Engineer Exam Questions and Answers by camila

Page: 12 / 16

Google Professional-Data-Engineer Exam Overview :

Exam Name: Google Professional Data Engineer Exam
Exam Code: Professional-Data-Engineer Dumps
Vendor: Google Certification: Google Cloud Certified
Questions: 387 Q&A's Shared By: camila
Question 48

You are implementing several batch jobs that must be executed on a schedule. These jobs have many interdependent steps that must be executed in a specific order. Portions of the jobs involve executing shell scripts, running Hadoop jobs, and running queries in BigQuery. The jobs are expected to run for many minutes up to several hours. If the steps fail, they must be retried a fixed number of times. Which service should you use to manage the execution of these jobs?

Options:

A.

Cloud Scheduler

B.

Cloud Dataflow

C.

Cloud Functions

D.

Cloud Composer

Discussion
Question 49

You want to schedule a number of sequential load and transformation jobs Data files will be added to a Cloud Storage bucket by an upstream process There is no fixed schedule for when the new data arrives Next, a Dataproc job is triggered to perform some transformations and write the data to BigQuery. You then need to run additional transformation jobs in BigQuery The transformation jobs are different for every table These jobs might take hours to complete You need to determine the most efficient and maintainable workflow to process hundreds of tables and provide the freshest data to your end users. What should you do?

Options:

A.

1Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Cloud Storage. Dataproc. and BigQuery operators2 Use a single shared DAG for all tables that need to go through the pipeline3 Schedule the DAG to run hourly

B.

1 Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Dataproc and BigQuery operators.2 Create a separate DAG for each table that needs to go through the pipeline3 Use a Cloud Storage object trigger to launch a Cloud Function that triggers the DAG

C.

1 Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Cloud Storage, Dataproc. and BigQuery operators2 Create a separate DAG for each table that needs to go through the pipeline3 Schedule the DAGs to run hourly

D.

1 Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Dataproc and BigQuery operators2 Use a single shared DAG for all tables that need to go through the pipeline.3 Use a Cloud Storage object trigger to launch a Cloud Function that triggers the DAG

Discussion
Question 50

You used Cloud Dataprep to create a recipe on a sample of data in a BigQuery table. You want to reuse this recipe on a daily upload of data with the same schema, after the load job with variable execution time completes. What should you do?

Options:

A.

Create a cron schedule in Cloud Dataprep.

B.

Create an App Engine cron job to schedule the execution of the Cloud Dataprep job.

C.

Export the recipe as a Cloud Dataprep template, and create a job in Cloud Scheduler.

D.

Export the Cloud Dataprep job as a Cloud Dataflow template, and incorporate it into a Cloud Composer job.

Discussion
Question 51

You are designing a messaging system by using Pub/Sub to process clickstream data with an event-driven consumer app that relies on a push subscription. You need to configure the messaging system that is reliable enough to handle temporary downtime of the consumer app. You also need the messaging system to store the input messages that cannot be consumed by the subscriber. The system needs to retry failed messages gradually, avoiding overloading the consumer app, and store the failed messages after a maximum of 10 retries in a topic. How should you configure the Pub/Sub subscription?

Options:

A.

Increase the acknowledgement deadline to 10 minutes.

B.

Use immediate redelivery as the subscription retry policy, and configure dead lettering to a different topic with maximum delivery attempts set to 10.

C.

Use exponential backoff as the subscription retry policy, and configure dead lettering to the same source topic with maximum delivery attempts set to 10.

D.

Use exponential backoff as the subscription retry policy, and configure dead lettering to a different topic with maximum delivery attempts set to 10.

Discussion
Anya
I must say they're considered the best dumps available and the questions are very similar to what you'll see in the actual exam. Recommended!!!
Cassius Dec 28, 2025
Yes, they offer a 100% success guarantee. And many students who have used them have reported passing their exams with flying colors.
Cecilia
Yes, I passed my certification exam using Cramkey Dumps.
Helena Dec 19, 2025
Great. Yes they are really effective
Miley
Hey, I tried Cramkey Dumps for my IT certification exam. They are really awesome and helped me pass my exam with wonderful score.
Megan Dec 16, 2025
That’s great!!! I’ll definitely give it a try. Thanks!!!
Elise
I've heard that Cramkey is one of the best websites for exam dumps. They have a high passing rate and the questions are always up-to-date. Is it true?
Cian Dec 27, 2025
Definitely. The dumps are constantly updated to reflect the latest changes in the certification exams. And I also appreciate how they provide explanations for the answers, so I could understand the reasoning behind each question.
Osian
Dumps are fantastic! I recently passed my certification exam using these dumps and I must say, they are 100% valid.
Azaan Dec 6, 2025
They are incredibly accurate and valid. I felt confident going into my exam because the dumps covered all the important topics and the questions were very similar to what I saw on the actual exam. The team of experts behind Cramkey Dumps make sure the information is relevant and up-to-date.
Page: 12 / 16
Title
Questions
Posted

Professional-Data-Engineer
PDF

$26.25  $104.99

Professional-Data-Engineer Testing Engine

$31.25  $124.99

Professional-Data-Engineer PDF + Testing Engine

$41.25  $164.99