Winter Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: dumps65

Google Professional-Cloud-Database-Engineer Dumps

Google Cloud Certified - Professional Cloud Database Engineer Questions and Answers

Question 1

You are starting a large CSV import into a Cloud SQL for MySQL instance that has many open connections. You checked memory and CPU usage, and sufficient resources are available. You want to follow Google-recommended practices to ensure that the import will not time out. What should you do?

Options:

A.

Close idle connections or restart the instance before beginning the import operation.

B.

Increase the amount of memory allocated to your instance.

C.

Ensure that the service account has the Storage Admin role.

D.

Increase the number of CPUs for the instance to ensure that it can handle the additional import operation.

Question 2

Your company wants to migrate an Oracle-based application to Google Cloud. The application team currently uses Oracle Recovery Manager (RMAN) to back up the database to tape for long-term retention (LTR). You need a cost-effective backup and restore solution that meets a 2-hour recovery time objective (RTO) and a 15-minute recovery point objective (RPO). What should you do?

Options:

A.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and store backups on tapes on-premises.

B.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and use Actifio to store backup files on Cloud Storage using the Nearline Storage class.

C.

Migrate the Oracle databases to Bare Metal Solution for Oracle, and back up the Oracle databases to Cloud Storage using the Standard Storage class.

D.

Migrate the Oracle databases to Compute Engine, and store backups on tapes on-premises.

Question 3

You are using Compute Engine on Google Cloud and your data center to manage a set of MySQL databases in a hybrid configuration. You need to create replicas to scale reads and to offload part of the management operation. What should you do?

Options:

A.

Use external server replication.

B.

Use Data Migration Service.

C.

Use Cloud SQL for MySQL external replica.

D.

Use the mysqldump utility and binary logs.

Question 4

Your team is building a new inventory management application that will require read and write database instances in multiple Google Cloud regions around the globe. Your database solution requires 99.99% availability and global transactional consistency. You need a fully managed backend relational database to store inventory changes. What should you do?

Options:

A.

Use Bigtable.

B.

Use Firestore.

C.

Use Cloud SQL for MySQL

D.

Use Cloud Spanner.

Question 5

Your company uses Cloud Spanner for a mission-critical inventory management system that is globally available. You recently loaded stock keeping unit (SKU) and product catalog data from a company acquisition and observed hot-spots in the Cloud Spanner database. You want to follow Google-recommended schema design practices to avoid performance degradation. What should you do? (Choose two.)

Options:

A.

Use an auto-incrementing value as the primary key.

B.

Normalize the data model.

C.

Promote low-cardinality attributes in multi-attribute primary keys.

D.

Promote high-cardinality attributes in multi-attribute primary keys.

E.

Use bit-reverse sequential value as the primary key.

Question 6

You are evaluating Cloud SQL for PostgreSQL as a possible destination for your on-premises PostgreSQL instances. Geography is becoming increasingly relevant to customer privacy worldwide. Your solution must support data residency requirements and include a strategy to:

configure where data is stored

control where the encryption keys are stored

govern the access to data

What should you do?

Options:

A.

Replicate Cloud SQL databases across different zones.

B.

Create a Cloud SQL for PostgreSQL instance on Google Cloud for the data that does not need to adhere to data residency requirements. Keep the data that must adhere to data residency requirements on-premises. Make application changes to support both databases.

C.

Allow application access to data only if the users are in the same region as the Google Cloud region for the Cloud SQL for PostgreSQL database.

D.

Use features like customer-managed encryption keys (CMEK), VPC Service Controls, and Identity and Access Management (IAM) policies.

Question 7

You need to provision several hundred Cloud SQL for MySQL instances for multiple project teams over a one-week period. You must ensure that all instances adhere to company standards such as instance naming conventions, database flags, and tags. What should you do?

Options:

A.

Automate instance creation by writing a Dataflow job.

B.

Automate instance creation by setting up Terraform scripts.

C.

Create the instances using the Google Cloud Console UI.

D.

Create clones from a template Cloud SQL instance.

Question 8

You work in the logistics department. Your data analysis team needs daily extracts from Cloud SQL for MySQL to train a machine learning model. The model will be used to optimize next-day routes. You need to export the data in CSV format. You want to follow Google-recommended practices. What should you do?

Options:

A.

Use Cloud Scheduler to trigger a Cloud Function that will run a select * from table(s) query to call the cloudsql.instances.export API.

B.

Use Cloud Scheduler to trigger a Cloud Function through Pub/Sub to call the cloudsql.instances.export API.

C.

Use Cloud Composer to orchestrate an export by calling the cloudsql.instances.export API.

D.

Use Cloud Composer to execute a select * from table(s) query and export results.

Question 9

Your organization is running a low-latency reporting application on Microsoft SQL Server. In addition to the database engine, you are using SQL Server Analysis Services (SSAS), SQL Server Reporting Services (SSRS), and SQL Server Integration Services (SSIS) in your on-premises environment. You want to migrate your Microsoft SQL Server database instances to Google Cloud. You need to ensure minimal disruption to the existing architecture during migration. What should you do?

Options:

A.

Migrate to Cloud SQL for SQL Server.

B.

Migrate to Cloud SQL for PostgreSQL.

C.

Migrate to Compute Engine.

D.

Migrate to Google Kubernetes Engine (GKE).

Question 10

You support a consumer inventory application that runs on a multi-region instance of Cloud Spanner. A customer opened a support ticket to complain about slow response times. You notice a Cloud Monitoring alert about high CPU utilization. You want to follow Google-recommended practices to address the CPU performance issue. What should you do first?

Options:

A.

Increase the number of processing units.

B.

Modify the database schema, and add additional indexes.

C.

Shard data required by the application into multiple instances.

D.

Decrease the number of processing units.

Question 11

You are managing a Cloud SQL for MySQL environment in Google Cloud. You have deployed a primary instance in Zone A and a read replica instance in Zone B, both in the same region. You are notified that the replica instance in Zone B was unavailable for 10 minutes. You need to ensure that the read replica instance is still working. What should you do?

Options:

A.

Use the Google Cloud Console or gcloud CLI to manually create a new clone database.

B.

Use the Google Cloud Console or gcloud CLI to manually create a new failover replica from backup.

C.

Verify that the new replica is created automatically.

D.

Start the original primary instance and resume replication.

Question 12

You need to migrate a 1 TB PostgreSQL database from a Compute Engine VM to Cloud SQL for PostgreSQL. You want to ensure that there is minimal downtime during the migration. What should you do?

Options:

A.

Export the data from the existing database, and load the data into a new Cloud SQL database.

B.

Use Migrate for Compute Engine to complete the migration.

C.

Use Datastream to complete the migration.

D.

Use Database Migration Service to complete the migration.

Question 13

Your organization is running a critical production database on a virtual machine (VM) on Compute Engine. The VM has an ext4-formatted persistent disk for data files. The database will soon run out of storage space. You need to implement a solution that avoids downtime. What should you do?

Options:

A.

In the Google Cloud Console, increase the size of the persistent disk, and use the resize2fs command to extend the disk.

B.

In the Google Cloud Console, increase the size of the persistent disk, and use the fdisk command to verify that the new space is ready to use

C.

In the Google Cloud Console, create a snapshot of the persistent disk, restore the snapshot to a new larger disk, unmount the old disk, mount the new disk, and restart the database service.

D.

In the Google Cloud Console, create a new persistent disk attached to the VM, and configure the database service to move the files to the new disk.

Question 14

You manage a production MySQL database running on Cloud SQL at a retail company. You perform routine maintenance on Sunday at midnight when traffic is slow, but you want to skip routine maintenance during the year-end holiday shopping season. You need to ensure that your production system is available 24/7 during the holidays. What should you do?

Options:

A.

Define a maintenance window on Sundays between 12 AM and 1 AM, and deny maintenance periods between November 1 and January 15.

B.

Define a maintenance window on Sundays between 12 AM and 5 AM, and deny maintenance periods between November 1 and February 15.

C.

Build a Cloud Composer job to start a maintenance window on Sundays between 12 AM and 1AM, and deny maintenance periods between November 1 and January 15.

D.

Create a Cloud Scheduler job to start maintenance at 12 AM on Sundays. Pause the Cloud Scheduler job between November 1 and January 15.

Question 15

You are working on a new centralized inventory management system to track items available in 200 stores, which each have 500 GB of data. You are planning a gradual rollout of the system to a few stores each week. You need to design an SQL database architecture that minimizes costs and user disruption during each regional rollout and can scale up or down on nights and holidays. What should you do?

Options:

A.

Use Oracle Real Application Cluster (RAC) databases on Bare Metal Solution for Oracle.

B.

Use sharded Cloud SQL instances with one or more stores per database instance.

C.

Use a Biglable cluster with autoscaling.

D.

Use Cloud Spanner with a custom autoscaling solution.

Question 16

You want to migrate an on-premises 100 TB Microsoft SQL Server database to Google Cloud over a 1 Gbps network link. You have 48 hours allowed downtime to migrate this database. What should you do? (Choose two.)

Options:

A.

Use a change data capture (CDC) migration strategy.

B.

Move the physical database servers from on-premises to Google Cloud.

C.

Keep the network bandwidth at 1 Gbps, and then perform an offline data migration.

D.

Increase the network bandwidth to 2 Gbps, and then perform an offline data migration.

E.

Increase the network bandwidth to 10 Gbps, and then perform an offline data migration.

Question 17

You are designing an augmented reality game for iOS and Android devices. You plan to use Cloud Spanner as the primary backend database for game state storage and player authentication. You want to track in-game rewards that players unlock at every stage of the game. During the testing phase, you discovered that costs are much higher than anticipated, but the query response times are within the SLA. You want to follow Google-recommended practices. You need the database to be performant and highly available while you keep costs low. What should you do?

Options:

A.

Manually scale down the number of nodes after the peak period has passed.

B.

Use interleaving to co-locate parent and child rows.

C.

Use the Cloud Spanner query optimizer to determine the most efficient way to execute the SQL query.

D.

Use granular instance sizing in Cloud Spanner and Autoscaler.

Question 18

Your customer has a global chat application that uses a multi-regional Cloud Spanner instance. The application has recently experienced degraded performance after a new version of the application was launched. Your customer asked you for assistance. During initial troubleshooting, you observed high read latency. What should you do?

Options:

A.

Use query parameters to speed up frequently executed queries.

B.

Change the Cloud Spanner configuration from multi-region to single region.

C.

Use SQL statements to analyze SPANNER_SYS.READ_STATS* tables.

D.

Use SQL statements to analyze SPANNER_SYS.QUERY_STATS* tables.

Question 19

Your company uses the Cloud SQL out-of-disk recommender to analyze the storage utilization trends of production databases over the last 30 days. Your database operations team uses these recommendations to proactively monitor storage utilization and implement corrective actions. You receive a recommendation that the instance is likely to run out of disk space. What should you do to address this storage alert?

Options:

A.

Normalize the database to the third normal form.

B.

Compress the data using a different compression algorithm.

C.

Manually or automatically increase the storage capacity.

D.

Create another schema to load older data.

Question 20

You want to migrate an on-premises mission-critical PostgreSQL database to Cloud SQL. The database must be able to withstand a zonal failure with less than five minutes of downtime and still not lose any transactions. You want to follow Google-recommended practices for the migration. What should you do?

Options:

A.

Take nightly snapshots of the primary database instance, and restore them in a secondary zone.

B.

Build a change data capture (CDC) pipeline to read transactions from the primary instance, and replicate them to a secondary instance.

C.

Create a read replica in another region, and promote the read replica if a failure occurs.

D.

Enable high availability (HA) for the database to make it regional.

Question 21

You are running a large, highly transactional application on Oracle Real Application Cluster (RAC) that is multi-tenant and uses shared storage. You need a solution that ensures high-performance throughput and a low-latency connection between applications and databases. The solution must also support existing Oracle features and provide ease of migration to Google Cloud. What should you do?

Options:

A.

Migrate to Compute Engine.

B.

Migrate to Bare Metal Solution for Oracle.

C.

Migrate to Google Kubernetes Engine (GKE)

D.

Migrate to Google Cloud VMware Engine

Question 22

You are running a mission-critical application on a Cloud SQL for PostgreSQL database with a multi-zonal setup. The primary and read replica instances are in the same region but in different zones. You need to ensure that you split the application load between both instances. What should you do?

Options:

A.

Use Cloud Load Balancing for load balancing between the Cloud SQL primary and read replica instances.

B.

Use PgBouncer to set up database connection pooling between the Cloud SQL primary and read replica instances.

C.

Use HTTP(S) Load Balancing for database connection pooling between the Cloud SQL primary and read replica instances.

D.

Use the Cloud SQL Auth proxy for database connection pooling between the Cloud SQL primary and read replica instances.

Question 23

An analytics team needs to read data out of Cloud SQL for SQL Server and update a table in Cloud Spanner. You need to create a service account and grant least privilege access using predefined roles. What roles should you assign to the service account?

Options:

A.

roles/cloudsql.viewer and roles/spanner.databaseUser

B.

roles/cloudsql.editor and roles/spanner.admin

C.

roles/cloudsql.client and roles/spanner.databaseReader

D.

roles/cloudsql.instanceUser and roles/spanner.databaseUser

Question 24

Your organization is currently updating an existing corporate application that is running in another public cloud to access managed database services in Google Cloud. The application will remain in the other public cloud while the database is migrated to Google Cloud. You want to follow Google-recommended practices for authentication. You need to minimize user disruption during the migration. What should you do?

Options:

A.

Use workload identity federation to impersonate a service account.

B.

Ask existing users to set their Google password to match their corporate password.

C.

Migrate the application to Google Cloud, and use Identity and Access Management (IAM).

D.

Use Google Workspace Password Sync to replicate passwords into Google Cloud.

Question 25

You are writing an application that will run on Cloud Run and require a database running in the Cloud SQL managed service. You want to secure this instance so that it only receives connections from applications running in your VPC environment in Google Cloud. What should you do?

Options:

A.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Use Cloud SQL Auth proxy to connect to the instance.

B.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Connect to the instance using a connection pool to best manage connections to the instance.

C.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Use Cloud SQL Auth proxy to connect to the instance.

D.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Connect to the instance using a connection pool to best manage connections to the instance.

Question 26

You are configuring a brand new PostgreSQL database instance in Cloud SQL. Your application team wants to have an optimal and highly available environment with automatic failover to avoid any unplanned outage. What should you do?

Options:

A.

Create one regional Cloud SQL instance with a read replica in another region.

B.

Create one regional Cloud SQL instance in one zone with a standby instance in another zone in the same region.

C.

Create two read-write Cloud SQL instances in two different zones with a standby instance in another region.

D.

Create two read-write Cloud SQL instances in two different regions with a standby instance in another zone.

Question 27

You are designing a highly available (HA) Cloud SQL for PostgreSQL instance that will be used by 100 databases. Each database contains 80 tables that were migrated from your on-premises environment to Google Cloud. The applications that use these databases are located in multiple regions in the US, and you need to ensure that read and write operations have low latency. What should you do?

Options:

A.

Deploy 2 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-east1 and us-west1.

B.

Deploy 2 Cloud SQL instances in the us-central1 region, and create read replicas in us-east1 and us-west1.

C.

Deploy 4 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-central1, us-east1, and us-west1.

D.

Deploy 4 Cloud SQL instances in the us-central1 region, and create read replicas in us-central1, us-east1 and us-west1.

Question 28

You are a DBA on a Cloud Spanner instance with multiple databases. You need to assign these privileges to all members of the application development team on a specific database:

Can read tables, views, and DDL

Can write rows to the tables

Can add columns and indexes

Cannot drop the database

What should you do?

Options:

A.

Assign the Cloud Spanner Database Reader and Cloud Spanner Backup Writer roles.

B.

Assign the Cloud Spanner Database Admin role.

C.

Assign the Cloud Spanner Database User role.

D.

Assign the Cloud Spanner Admin role.

Question 29

You are troubleshooting a connection issue with a newly deployed Cloud SQL instance on Google Cloud. While investigating the Cloud SQL Proxy logs, you see the message Error 403: Access Not Configured. What should you do?

Options:

A.

Check the app.yaml value cloud_sql_instances for a misspelled or incorrect instance connection name.

B.

Check whether your service account has cloudsql.instances.connect permission.

C.

Enable the Cloud SQL Admin API.

D.

Ensure that you are using an external (public) IP address interface.

Question 30

Your organization is migrating 50 TB Oracle databases to Bare Metal Solution for Oracle. Database backups must be available for quick restore. You also need to have backups available for 5 years. You need to design a cost-effective architecture that meets a recovery time objective (RTO) of 2 hours and recovery point objective (RPO) of 15 minutes. What should you do?

Options:

A.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on all flash storage.

Keep backups older than one day stored in Actifio OnVault storage.

B.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on standard storage.

Keep backups older than one day stored in Actifio OnVault storage.

C.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on standard storage.

Use the Oracle Recovery Manager (RMAN) backup utility to move backups older than one day to a Coldline Storage bucket.

D.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on all flash storage.

Use the Oracle Recovery Manager (RMAN) backup utility to move backups older than one day to an Archive Storage bucket.

Question 31

Your organization works with sensitive data that requires you to manage your own encryption keys. You are working on a project that stores that data in a Cloud SQL database. You need to ensure that stored data is encrypted with your keys. What should you do?

Options:

A.

Export data periodically to a Cloud Storage bucket protected by Customer-Supplied Encryption Keys.

B.

Use Cloud SQL Auth proxy.

C.

Connect to Cloud SQL using a connection that has SSL encryption.

D.

Use customer-managed encryption keys with Cloud SQL.

Question 32

You have deployed a Cloud SQL for SQL Server instance. In addition, you created a cross-region read replica for disaster recovery (DR) purposes. Your company requires you to maintain and monitor a recovery point objective (RPO) of less than 5 minutes. You need to verify that your cross-region read replica meets the allowed RPO. What should you do?

Options:

A.

Use Cloud SQL instance monitoring.

B.

Use the Cloud Monitoring dashboard with available metrics from Cloud SQL.

C.

Use Cloud SQL logs.

D.

Use the SQL Server Always On Availability Group dashboard.

Question 33

You plan to use Database Migration Service to migrate data from a PostgreSQL on-premises instance to Cloud SQL. You need to identify the prerequisites for creating and automating the task. What should you do? (Choose two.)

Options:

A.

Drop or disable all users except database administration users.

B.

Disable all foreign key constraints on the source PostgreSQL database.

C.

Ensure that all PostgreSQL tables have a primary key.

D.

Shut down the database before the Data Migration Service task is started.

E.

Ensure that pglogical is installed on the source PostgreSQL database.

Question 34

You released a popular mobile game and are using a 50 TB Cloud Spanner instance to store game data in a PITR-enabled production environment. When you analyzed the game statistics, you realized that some players are exploiting a loophole to gather more points to get on the leaderboard. Another DBA accidentally ran an emergency bugfix script that corrupted some of the data in the production environment. You need to determine the extent of the data corruption and restore the production environment. What should you do? (Choose two.)

Options:

A.

If the corruption is significant, use backup and restore, and specify a recovery timestamp.

B.

If the corruption is significant, perform a stale read and specify a recovery timestamp. Write the results back.

C.

If the corruption is significant, use import and export.

D.

If the corruption is insignificant, use backup and restore, and specify a recovery timestamp.

E.

If the corruption is insignificant, perform a stale read and specify a recovery timestamp. Write the results back.

Question 35

You are migrating an on-premises application to Compute Engine and Cloud SQL. The application VMs will live in their own project, separate from the Cloud SQL instances which have their own project. What should you do to configure the networks?

Options:

A.

Create a new VPC network in each project, and use VPC Network Peering to connect the two together.

B.

Create a Shared VPC that both the application VMs and Cloud SQL instances will use.

C.

Use the default networks, and leverage Cloud VPN to connect the two together.

D.

Place both the application VMs and the Cloud SQL instances in the default network of each project.

Question 36

You are configuring a brand new Cloud SQL for PostgreSQL database instance in Google Cloud. Your application team wants you to deploy one primary instance, one standby instance, and one read replica instance. You need to ensure that you are following Google-recommended practices for high availability. What should you do?

Options:

A.

Configure the primary instance in zone A, the standby instance in zone C, and the read replica in zone B, all in the same region.

B.

Configure the primary and standby instances in zone A and the read replica in zone B, all in the same region.

C.

Configure the primary instance in one region, the standby instance in a second region, and the read replica in a third region.

D.

Configure the primary, standby, and read replica instances in zone A, all in the same region.

Question 37

Your company uses Bigtable for a user-facing application that displays a low-latency real-time dashboard. You need to recommend the optimal storage type for this read-intensive database. What should you do?

Options:

A.

Recommend solid-state drives (SSD).

B.

Recommend splitting the Bigtable instance into two instances in order to load balance the concurrent reads.

C.

Recommend hard disk drives (HDD).

D.

Recommend mixed storage types.

Question 38

You are designing a physician portal app in Node.js. This application will be used in hospitals and clinics that might have intermittent internet connectivity. If a connectivity failure occurs, the app should be able to query the cached data. You need to ensure that the application has scalability, strong consistency, and multi-region replication. What should you do?

Options:

A.

Use Firestore and ensure that the PersistenceEnabled option is set to true.

B.

Use Memorystore for Memcached.

C.

Use Pub/Sub to synchronize the changes from the application to Cloud Spanner.

D.

Use Table.read with the exactStaleness option to perform a read of rows in Cloud Spanner.

Question 39

You are managing multiple applications connecting to a database on Cloud SQL for PostgreSQL. You need to be able to monitor database performance to easily identify applications with long-running and resource-intensive queries. What should you do?

Options:

A.

Use log messages produced by Cloud SQL.

B.

Use Query Insights for Cloud SQL.

C.

Use the Cloud Monitoring dashboard with available metrics from Cloud SQL.

D.

Use Cloud SQL instance monitoring in the Google Cloud Console.

Page: 1 / 13
Total 132 questions