Summer Special - 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: c4sdisc65

Professional-Cloud-Database-Engineer PDF

$38.5

$109.99

3 Months Free Update

  • Printable Format
  • Value of Money
  • 100% Pass Assurance
  • Verified Answers
  • Researched by Industry Experts
  • Based on Real Exams Scenarios
  • 100% Real Questions

Professional-Cloud-Database-Engineer PDF + Testing Engine

$61.6

$175.99

3 Months Free Update

  • Exam Name: Google Cloud Certified - Professional Cloud Database Engineer
  • Last Update: Sep 12, 2025
  • Questions and Answers: 141
  • Free Real Questions Demo
  • Recommended by Industry Experts
  • Best Economical Package
  • Immediate Access

Professional-Cloud-Database-Engineer Engine

$46.2

$131.99

3 Months Free Update

  • Best Testing Engine
  • One Click installation
  • Recommended by Teachers
  • Easy to use
  • 3 Modes of Learning
  • State of Art Technology
  • 100% Real Questions included

Professional-Cloud-Database-Engineer Practice Exam Questions with Answers Google Cloud Certified - Professional Cloud Database Engineer Certification

Question # 6

You are the database administrator of a Cloud SQL for PostgreSQL instance that has pgaudit disabled. Users are complaining that their queries are taking longer to execute and performance has degraded over the past few months. You need to collect and analyze query performance data to help identity slow-running queries. What should you do?

A.

View Cloud SQL operations to view historical query information.

B.

White a Logs Explorer query to identify database queries with high execution times.

C.

Review application logs to identify database calls.

D.

Use the Query Insights dashboard to identify high execution times.

Full Access
Question # 7

Your team is building an application that stores and analyzes streaming time series financial data. You need a database solution that can perform time series-based scans with sub-second latency. The solution must scale into the hundreds of terabytes and be able to write up to 10k records per second and read up to 200 MB per second. What should you do?

A.

Use Firestore.

B.

Use Bigtable

C.

Use BigQuery.

D.

Use Cloud Spanner.

Full Access
Question # 8

Your team recently released a new version of a highly consumed application to accommodate additional user traffic. Shortly after the release, you received an alert from your production monitoring team that there is consistently high replication lag between your primary instance and the read replicas of your Cloud SQL for MySQL instances. You need to resolve the replication lag. What should you do?

A.

Identify and optimize slow running queries, or set parallel replication flags.

B.

Stop all running queries, and re-create the replicas.

C.

Edit the primary instance to upgrade to a larger disk, and increase vCPU count.

D.

Edit the primary instance to add additional memory.

Full Access
Question # 9

Your application follows a microservices architecture and uses a single large Cloud SQL instance, which is starting to have performance issues as your application grows. in the Cloud Monitoring dashboard, the CPU utilization looks normal You want to followGoogle-recommended practices to resolve and prevent these performance issues while avoiding any major refactoring. What should you do?

A.

Use Cloud Spanner instead of Cloud SQL.

B.

Increase the number of CPUs for your instance.

C.

Increase the storage size for the instance.

D.

Use many smaller Cloud SQL instances.

Full Access
Question # 10

You are developing a new application on a VM that is on your corporate network. The application will use Java Database Connectivity (JDBC) to connect to Cloud SQL for PostgreSQL. Your Cloud SQL instance is configured with IP address 192.168.3.48, and SSL is disabled. You want to ensure that your application can access your database instance without requiring configuration changes to your database. What should you do?

A.

Define a connection string using your Google username and password to point to the external (public) IP address of your Cloud SQL instance.

B.

Define a connection string using a database username and password to point to the internal (private) IP address of your Cloud SQL instance.

C.

Define a connection string using Cloud SQL Auth proxy configured with a service account to point to the internal (private) IP address of your Cloud SQL instance.

D.

Define a connection string using Cloud SQL Auth proxy configured with a service account to point to the external (public) IP address of your Cloud SQL instance.

Full Access
Question # 11

Your company is evaluating Google Cloud database options for a mission-critical global payments gateway application. The application must be available 24/7 to users worldwide, horizontally scalable, and support open source databases. You need to select an automatically shardable, fully managed database with 99.999% availability and strong transactional consistency. What should you do?

A.

Select Bare Metal Solution for Oracle.

B.

Select Cloud SQL.

C.

Select Bigtable.

D.

Select Cloud Spanner.

Full Access
Question # 12

Your team is running a Cloud SQL for MySQL instance with a 5 TB database that must be available 24/7. You need to save database backups on object storage with minimal operational overhead or risk to your production workloads. What should you do?

A.

Use Cloud SQL serverless exports.

B.

Create a read replica, and then use the mysqldump utility to export each table.

C.

Clone the Cloud SQL instance, and then use the mysqldump utlity to export the data.

D.

Use the mysqldump utility on the primary database instance to export the backup.

Full Access
Question # 13

Your online delivery business that primarily serves retail customers uses Cloud SQL for MySQL for its inventory and scheduling application. The required recovery time objective (RTO) and recovery point objective (RPO) must be in minutes rather than hours as a part of your high availability and disaster recovery design. You need a high availability configuration that can recover without data loss during a zonal or a regional failure. What should you do?

A.

Set up all read replicas in a different region using asynchronous replication.

B.

Set up all read replicas in the same region as the primary instance with synchronous replication.

C.

Set up read replicas in different zones of the same region as the primary instance with synchronous replication, and set up read replicas in different regions with asynchronous replication.

D.

Set up read replicas in different zones of the same region as the primary instance with asynchronous replication, and set up read replicas in different regions with synchronous replication.

Full Access
Question # 14

Your company has PostgreSQL databases on-premises and on Amazon Web Services (AWS). You are planning multiple database migrations to Cloud SQL in an effort to reduce costs and downtime. You want to follow Google-recommended practices anduse Google native data migration tools. You also want to closely monitor the migrations as part of the cutover strategy. What should you do?

A.

Use Database Migration Service to migrate all databases to Cloud SQL.

B.

Use Database Migration Service for one-time migrations, and use third-party or partner tools for change data capture (CDC) style migrations.

C.

Use data replication tools and CDC tools to enable migration.

D.

Use a combination of Database Migration Service and partner tools to support the data migration strategy.

Full Access
Question # 15

You are running a mission-critical application on a Cloud SQL for PostgreSQL database with a multi-zonal setup. The primary and read replica instances are in the same regionbut in different zones. You need to ensure that you split the application load between both instances. What should you do?

A.

Use Cloud Load Balancing for load balancing between the Cloud SQL primary and read replica instances.

B.

Use PgBouncer to set up database connection pooling between the Cloud SQL primary and read replica instances.

C.

Use HTTP(S) Load Balancing for database connection pooling between the Cloud SQL primary and read replica instances.

D.

Use the Cloud SQL Auth proxy for database connection pooling between the Cloud SQL primary and read replica instances.

Full Access
Question # 16

You are migrating an on-premises application to Google Cloud. The application requires a high availability (HA) PostgreSQL database to support business-critical functions. Your company's disaster recovery strategy requires a recovery time objective (RTO) and recovery point objective (RPO) within 30 minutes of failure. You plan to use a Google Cloud managed service. What should you do to maximize uptime for your application?

A.

Deploy Cloud SQL for PostgreSQL in a regional configuration. Create a read replica in a different zone in the same region and a read replica in another region for disaster recovery.

B.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Take periodic backups, and use this backup to restore to a new Cloud SQL for PostgreSQL instance in another region during a disaster recovery event.

C.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Create a cross-region read replica, and promote the read replica as the primary node for disaster recovery.

D.

Migrate the PostgreSQL database to multi-regional Cloud Spanner so that a single region outage will not affect your application. Update the schema to support Cloud Spanner data types, and refactor the application.

Full Access
Question # 17

You are choosing a new database backend for an existing application. The current database is running PostgreSQL on an on-premises VM and is managed by a database administrator and operations team. The application data is relational and has light traffic. You want to minimize costs and the migration effort for this application. What should you do?

A.

Migrate the existing database to Firestore.

B.

Migrate the existing database to Cloud SQL for PostgreSQL.

C.

Migrate the existing database to Cloud Spanner.

D.

Migrate the existing database to PostgreSQL running on Compute Engine.

Full Access
Question # 18

Your organization has a ticketing system that needs an online marketing analytics and reporting application. You need to select a relational database that can manage hundreds of terabytes of data to support this new application. Which database should you use?

A.

Cloud SQL

B.

BigQuery

C.

Cloud Spanner

D.

Bigtable

Full Access
Question # 19

You recently launched a new product to the US market. You currently have two Bigtable clusters in one US region to serve all the traffic. Your marketing team is planning an immediate expansion to APAC. You need to roll out the regional expansion while implementing high availability according to Google-recommended practices. What should you do?

A.

Maintain a target of 23% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone europe-west1-d

cluster-c in zone asia-east1-b

B.

Maintain a target of 23% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone us-central1-b

cluster-c in zone us-east1-a

C.

Maintain a target of 35% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone australia-southeast1-a

cluster-c in zone europe-west1-d

cluster-d in zone asia-east1-b

D.

Maintain a target of 35% CPU utilization by locating:

cluster-a in zone us-central1-a

cluster-b in zone us-central2-a

cluster-c in zone asia-northeast1-b

cluster-d in zone asia-east1-b

Full Access
Question # 20

You are troubleshooting a connection issue with a newly deployed Cloud SQL instance on Google Cloud. While investigating the Cloud SQL Proxy logs, you see the message Error 403: Access Not Configured. What should you do?

A.

Check the app.yaml value cloud_sql_instances for a misspelled or incorrect instance connection name.

B.

Check whether your service account has cloudsql.instances.connect permission.

C.

Enable the Cloud SQL Admin API.

D.

Ensure that you are using an external (public) IP address interface.

Full Access
Question # 21

Your company is developing a new global transactional application that must be ACID-compliant and have 99.999% availability. You are responsible for selecting the appropriate Google Cloud database to serve as a datastore for this new application. What should you do?

A.

Use Firestore.

B.

Use Cloud Spanner.

C.

Use Cloud SQL.

D.

Use Bigtable.

Full Access
Question # 22

Your team is building a new inventory management application that will require read and write database instances in multiple Google Cloud regions around the globe. Your database solution requires 99.99% availability and global transactional consistency. You need a fully managed backend relational database to store inventory changes. What should you do?

A.

Use Bigtable.

B.

Use Firestore.

C.

Use Cloud SQL for MySQL

D.

Use Cloud Spanner.

Full Access
Question # 23

You have a Cloud SQL instance (DB-1) with two cross-region read replicas (DB-2 and DB-3). During a business continuity test, the primary instance (DB-1) was taken offline and a replica (DB-2) was promoted. The test has concluded and you want to return to the pre-test configuration. What should you do?

A.

Bring DB-1 back online.

B.

Delete DB-1, and re-create DB-1 as a read replica in the same region as DB-1.

C.

Delete DB-2 so that DB-1 automatically reverts to the primary instance.

D.

Create DB-4 as a read replica in the same region as DB-1, and promote DB-4 to primary.

Full Access
Question # 24

Your company is migrating all legacy applications to Google Cloud. All on-premises applications are using legacy Oracle 12c databases with Oracle Real Application Cluster (RAC) for high availability (HA) and Oracle Data Guard for disaster recovery. You need a solution that requires minimal code changes, provides the same high availability you have today on-premises, and supports a low latency network for migrated legacy applications. What should you do?

A.

Migrate the databases to Cloud Spanner.

B.

Migrate the databases to Cloud SQL, and enable a standby database.

C.

Migrate the databases to Compute Engine using regional persistent disks.

D.

Migrate the databases to Bare Metal Solution for Oracle.

Full Access
Question # 25

Your organization has an existing app that just went viral. The app uses a Cloud SQL for MySQL backend database that is experiencing slow disk performance while using hard disk drives (HDDs). You need to improve performance and reduce disk I/O wait times. What should you do?

A.

Export the data from the existing instance, and import the data into a new instance with solid-state drives (SSDs).

B.

Edit the instance to change the storage type from HDD to SSD.

C.

Create a high availability (HA) failover instance with SSDs, and perform a failover to the new instance.

D.

Create a read replica of the instance with SSDs, and perform a failover to the new instance

Full Access
Question # 26

You are evaluating Cloud SQL for PostgreSQL as a possible destination for your on-premises PostgreSQL instances. Geography is becoming increasingly relevant to customer privacy worldwide. Your solution must support data residency requirements and include a strategy to:

configure where data is stored

control where the encryption keys are stored

govern the access to data

What should you do?

A.

Replicate Cloud SQL databases across different zones.

B.

Create a Cloud SQL for PostgreSQL instance on Google Cloud for the data that does not need to adhere to data residency requirements. Keep the data that must adhere to data residency requirements on-premises. Make application changes to support both databases.

C.

Allow application access to data only if the users are in the same region as the Google Cloud region for the Cloud SQL for PostgreSQL database.

D.

Use features like customer-managed encryption keys (CMEK), VPC Service Controls, and Identity and Access Management (IAM) policies.

Full Access
Question # 27

Your company is using Cloud SQL for MySQL with an internal (private) IP address and wants to replicate some tables into BigQuery in near-real time for analytics and machine learning. You need to ensure that replication is fast and reliable and uses Google-managed services. What should you do?

A.

Develop a custom data replication service to send data into BigQuery.

B.

Use Cloud SQL federated queries.

C.

Use Database Migration Service to replicate tables into BigQuery.

D.

Use Datastream to capture changes, and use Dataflow to write those changes to BigQuery.

Full Access
Question # 28

You are configuring a new application that has access to an existing Cloud Spanner database. The new application reads from this database to gather statistics for a dashboard. You want to follow Google-recommended practices when granting Identity and Access Management (IAM) permissions. What should you do?

A.

Reuse the existing service account that populates this database.

B.

Create a new service account, and grant it the Cloud Spanner Database Admin role.

C.

Create a new service account, and grant it the Cloud Spanner Database Reader role.

D.

Create a new service account, and grant it the spanner.databases.select permission.

Full Access
Question # 29

You are managing a set of Cloud SQL databases in Google Cloud. Regulations require that database backups reside in the region where the database is created. You want to minimize operational costs and administrative effort. What should you do?

A.

Configure the automated backups to use a regional Cloud Storage bucket as a custom location.

B.

Use the default configuration for the automated backups location.

C.

Disable automated backups, and create an on-demand backup routine to a regional Cloud Storage bucket.

D.

Disable automated backups, and configure serverless exports to a regional Cloud Storage bucket.

Full Access
Question # 30

You want to migrate an on-premises 100 TB Microsoft SQL Server database to Google Cloud over a 1 Gbps network link. You have 48 hours allowed downtime to migrate this database. What should you do? (Choose two.)

A.

Use a change data capture (CDC) migration strategy.

B.

Move the physical database servers from on-premises to Google Cloud.

C.

Keep the network bandwidth at 1 Gbps, and then perform an offline data migration.

D.

Increase the network bandwidth to 2 Gbps, and then perform an offline data migration.

E.

Increase the network bandwidth to 10 Gbps, and then perform an offline data migration.

Full Access
Question # 31

Your customer has a global chat application that uses a multi-regional Cloud Spanner instance. The application has recently experienced degraded performance after a new version of the application was launched. Your customer asked you for assistance. During initial troubleshooting, you observed high read latency. What should you do?

A.

Use query parameters to speed up frequently executed queries.

B.

Change the Cloud Spanner configuration from multi-region to single region.

C.

Use SQL statements to analyze SPANNER_SYS.READ_STATS* tables.

D.

Use SQL statements to analyze SPANNER_SYS.QUERY_STATS* tables.

Full Access
Question # 32

You are the DBA of an online tutoring application that runs on a Cloud SQL for PostgreSQL database. You are testing the implementation of the cross-regional failover configuration. The database in region R1 fails over successfully to region R2, and the database becomes available for the application to process data. During testing, certain scenarios of the application work as expected in region R2, but a few scenarios fail with database errors. The application-related database queries, when executed in isolation from Cloud SQL for PostgreSQL in region R2, work as expected. The applicationperforms completely as expected when the database fails back to region R1. You need to identify the cause of the database errors in region R2. What should you do?

A.

Determine whether the versions of Cloud SQL for PostgreSQL in regions R1 and R2 are different.

B.

Determine whether the database patches of Cloud SQI for PostgreSQL in regions R1 and R2 are different.

C.

Determine whether the failover of Cloud SQL for PostgreSQL from region R1 to region R2 is in progress or has completed successfully.

D.

Determine whether Cloud SQL for PostgreSQL in region R2 is a near-real-time copy of region R1 but not an exact copy.

Full Access
Question # 33

You want to migrate your on-premises PostgreSQL database to Compute Engine. You need to migrate this database with the minimum downtime possible. What should you do?

A.

Perform a full backup of your on-premises PostgreSQL, and then, in the migration window, perform an incremental backup.

B.

Create a read replica on Cloud SQL, and then promote it to a read/write standalone instance.

C.

Use Database Migration Service to migrate your database.

D.

Create a hot standby on Compute Engine, and use PgBouncer to switch over the connections.

Full Access
Question # 34

You are designing a database architecture for a global application that stores information about public parks worldwide. The application uses the database for read-only purposes, and a centralized batch job updates the database nightly. You want to select an open source, SQL-compliant database. What should you do?

A.

Use Bigtable with multi-region clusters.

B.

Use Memorystore for Redis with multi-zones within a region.

C.

Use Cloud SQL for PostgreSQL with cross-region replicas.

D.

Use Cloud Spanner with multi-region configuration.

Full Access
Question # 35

Your organization has a busy transactional Cloud SQL for MySQL instance. Your analytics team needs access to the data so they can build monthly sales reports. You need to provide data access to the analytics team without adversely affecting performance. What should you do?

A.

Create a read replica of the database, provide the database IP address, username, and password to the analytics team, and grant read access to required tables to the team.

B.

Create a read replica of the database, enable the cloudsql.iam_authentication flag on the replica, and grant read access to required tables to the analytics team.

C.

Enable the cloudsql.iam_authentication flag on the primary database instance, and grant read access to required tables to the analytics team.

D.

Provide the database IP address, username, and password of the primary database instance to the analytics, team, and grant read access to required tables to the team.

Full Access
Question # 36

Your organization is currently updating an existing corporate application that is running in another public cloud to access managed database services in Google Cloud. The application will remain in the other public cloud while the database is migrated to Google Cloud. You want to follow Google-recommended practices for authentication. You need to minimize user disruption during the migration. What should you do?

A.

Use workload identity federation to impersonate a service account.

B.

Ask existing users to set their Google password to match their corporate password.

C.

Migrate the application to Google Cloud, and use Identity and Access Management (IAM).

D.

Use Google Workspace Password Sync to replicate passwords into Google Cloud.

Full Access
Question # 37

You need to migrate existing databases from Microsoft SQL Server 2016 Standard Edition on a single Windows Server 2019 Datacenter Edition to a single Cloud SQL for SQL Server instance. During the discovery phase of your project, you notice that your on-premises server peaks at around 25,000 read IOPS. You need to ensure that your Cloud SQL instance is sized appropriately to maximize read performance. What should you do?

A.

Create a SQL Server 2019 Standard on Standard machine type with 4 vCPUs, 15 GB of RAM, and 800 GB of solid-state drive (SSD).

B.

Create a SQL Server 2019 Standard on High Memory machine type with at least 16 vCPUs, 104 GB of RAM, and 200 GB of SSD.

C.

Create a SQL Server 2019 Standard on High Memory machine type with 16 vCPUs, 104 GB of RAM, and 4 TB of SSD.

D.

Create a SQL Server 2019 Enterprise on High Memory machine type with 16 vCPUs, 104 GB of RAM, and 500 GB of SSD.

Full Access
Question # 38

You are writing an application that will run on Cloud Run and require a database running in the Cloud SQL managed service. You want to secure this instance so that it only receives connections from applications running in your VPC environment in Google Cloud. What should you do?

A.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Use Cloud SQL Auth proxy to connect to the instance.

B.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Connect to the instance using a connection pool to best manage connections to the instance.

C.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Use Cloud SQL Auth proxy to connect to the instance.

D.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Connect to the instance using a connection pool to best manage connections to the instance.

Full Access
Question # 39

You are designing a highly available (HA) Cloud SQL for PostgreSQL instance that will be used by 100 databases. Each database contains 80 tables that were migrated from your on-premises environment to Google Cloud. The applications that use these databases are located in multiple regions in the US, and you need to ensure that read and write operations have low latency. What should you do?

A.

Deploy 2 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-east1 and us-west1.

B.

Deploy 2 Cloud SQL instances in the us-central1 region, and create read replicas in us-east1 and us-west1.

C.

Deploy 4 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-central1, us-east1, and us-west1.

D.

Deploy 4 Cloud SQL instances in the us-central1 region, and create read replicas in us-central1, us-east1 and us-west1.

Full Access
Question # 40

You are designing a database strategy for a new web application in one region. You need to minimize write latency. What should you do?

A.

Use Cloud SQL with cross-region replicas.

B.

Use high availability (HA) Cloud SQL with multiple zones.

C.

Use zonal Cloud SQL without high availability (HA).

D.

Use Cloud Spanner in a regional configuration.

Full Access
Question # 41

You want to migrate an on-premises mission-critical PostgreSQL database to Cloud SQL. The database must be able to withstand a zonal failure with less than five minutes of downtime and still not lose any transactions. You want to follow Google-recommended practices for the migration. What should you do?

A.

Take nightly snapshots of the primary database instance, and restore them in a secondary zone.

B.

Build a change data capture (CDC) pipeline to read transactions from the primary instance, and replicate them to a secondary instance.

C.

Create a read replica in another region, and promote the read replica if a failure occurs.

D.

Enable high availability (HA) for the database to make it regional.

Full Access
Question # 42

You are a DBA of Cloud SQL for PostgreSQL. You want the applications to have password-less authentication for read and write access to the database. Which authentication mechanism should you use?

A.

Use Identity and Access Management (IAM) authentication.

B.

Use Managed Active Directory authentication.

C.

Use Cloud SQL federated queries.

D.

Use PostgreSQL database's built-in authentication.

Full Access