New Year Special 75% Discount offer - Ends in 0d 00h 00m 00s - Coupon code: 75brite

Google Updated Professional-Data-Engineer Exam Questions and Answers by camila

Page: 12 / 16

Google Professional-Data-Engineer Exam Overview :

Exam Name: Google Professional Data Engineer Exam
Exam Code: Professional-Data-Engineer Dumps
Vendor: Google Certification: Google Cloud Certified
Questions: 387 Q&A's Shared By: camila
Question 48

You are implementing several batch jobs that must be executed on a schedule. These jobs have many interdependent steps that must be executed in a specific order. Portions of the jobs involve executing shell scripts, running Hadoop jobs, and running queries in BigQuery. The jobs are expected to run for many minutes up to several hours. If the steps fail, they must be retried a fixed number of times. Which service should you use to manage the execution of these jobs?

Options:

A.

Cloud Scheduler

B.

Cloud Dataflow

C.

Cloud Functions

D.

Cloud Composer

Discussion
Question 49

You want to schedule a number of sequential load and transformation jobs Data files will be added to a Cloud Storage bucket by an upstream process There is no fixed schedule for when the new data arrives Next, a Dataproc job is triggered to perform some transformations and write the data to BigQuery. You then need to run additional transformation jobs in BigQuery The transformation jobs are different for every table These jobs might take hours to complete You need to determine the most efficient and maintainable workflow to process hundreds of tables and provide the freshest data to your end users. What should you do?

Options:

A.

1Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Cloud Storage. Dataproc. and BigQuery operators2 Use a single shared DAG for all tables that need to go through the pipeline3 Schedule the DAG to run hourly

B.

1 Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Dataproc and BigQuery operators.2 Create a separate DAG for each table that needs to go through the pipeline3 Use a Cloud Storage object trigger to launch a Cloud Function that triggers the DAG

C.

1 Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Cloud Storage, Dataproc. and BigQuery operators2 Create a separate DAG for each table that needs to go through the pipeline3 Schedule the DAGs to run hourly

D.

1 Create an Apache Airflow directed acyclic graph (DAG) in Cloud Composer with sequential tasks by using the Dataproc and BigQuery operators2 Use a single shared DAG for all tables that need to go through the pipeline.3 Use a Cloud Storage object trigger to launch a Cloud Function that triggers the DAG

Discussion
Question 50

You used Cloud Dataprep to create a recipe on a sample of data in a BigQuery table. You want to reuse this recipe on a daily upload of data with the same schema, after the load job with variable execution time completes. What should you do?

Options:

A.

Create a cron schedule in Cloud Dataprep.

B.

Create an App Engine cron job to schedule the execution of the Cloud Dataprep job.

C.

Export the recipe as a Cloud Dataprep template, and create a job in Cloud Scheduler.

D.

Export the Cloud Dataprep job as a Cloud Dataflow template, and incorporate it into a Cloud Composer job.

Discussion
Anaya
I found so many of the same questions on the real exam that I had already seen in the Cramkey Dumps. Thank you so much for making exam so easy for me. I passed it successfully!!!
Nina Dec 25, 2025
It's true! I felt so much more confident going into the exam because I had already seen and understood the questions.
Pippa
I was so happy to see that almost all the questions on the exam were exactly what I found in their Dumps.
Anastasia Dec 21, 2025
You are right…It was amazing! The Cramkey Dumps were so comprehensive and well-organized, it made studying for the exam a breeze.
Aliza
I used these dumps for my recent certification exam and I can say with certainty that they're absolutely valid dumps. The questions were very similar to what came up in the actual exam.
Jakub Dec 3, 2025
That's great to hear. I am going to try them soon.
Ernest
That's amazing. I think I'm going to give Cramkey Dumps a try for my next exam. Thanks for telling me about them! CramKey admin please share more questions……You guys are amazing.
Nate Dec 19, 2025
I failed last week, I never know this site , but amazed to see all these questions were in my exam week before. I feel bad now, why I didn’t bother this site. Thanks Cramkey, Excellent Job.
Cody
I used Cramkey Dumps to prepare and a lot of the questions on the exam were exactly what I found in their study materials.
Eric Dec 2, 2025
Really? That's great to hear! I used Cramkey Dumps too and I had the same experience. The questions were almost identical.
Question 51

You are designing a messaging system by using Pub/Sub to process clickstream data with an event-driven consumer app that relies on a push subscription. You need to configure the messaging system that is reliable enough to handle temporary downtime of the consumer app. You also need the messaging system to store the input messages that cannot be consumed by the subscriber. The system needs to retry failed messages gradually, avoiding overloading the consumer app, and store the failed messages after a maximum of 10 retries in a topic. How should you configure the Pub/Sub subscription?

Options:

A.

Increase the acknowledgement deadline to 10 minutes.

B.

Use immediate redelivery as the subscription retry policy, and configure dead lettering to a different topic with maximum delivery attempts set to 10.

C.

Use exponential backoff as the subscription retry policy, and configure dead lettering to the same source topic with maximum delivery attempts set to 10.

D.

Use exponential backoff as the subscription retry policy, and configure dead lettering to a different topic with maximum delivery attempts set to 10.

Discussion
Page: 12 / 16
Title
Questions
Posted

Professional-Data-Engineer
PDF

$26.25  $104.99

Professional-Data-Engineer Testing Engine

$31.25  $124.99

Professional-Data-Engineer PDF + Testing Engine

$41.25  $164.99