Labour Day - Special 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 70dumps

Professional-Cloud-Database-Engineer Questions and Answers

Question # 6

You are managing a set of Cloud SQL databases in Google Cloud. Regulations require that database backups reside in the region where the database is created. You want to minimize operational costs and administrative effort. What should you do?

A.

Configure the automated backups to use a regional Cloud Storage bucket as a custom location.

B.

Use the default configuration for the automated backups location.

C.

Disable automated backups, and create an on-demand backup routine to a regional Cloud Storage bucket.

D.

Disable automated backups, and configure serverless exports to a regional Cloud Storage bucket.

Full Access
Question # 7

Your application follows a microservices architecture and uses a single large Cloud SQL instance, which is starting to have performance issues as your application grows. in the Cloud Monitoring dashboard, the CPU utilization looks normal You want to follow Google-recommended practices to resolve and prevent these performance issues while avoiding any major refactoring. What should you do?

A.

Use Cloud Spanner instead of Cloud SQL.

B.

Increase the number of CPUs for your instance.

C.

Increase the storage size for the instance.

D.

Use many smaller Cloud SQL instances.

Full Access
Question # 8

You finished migrating an on-premises MySQL database to Cloud SQL. You want to ensure that the daily export of a table, which was previously a cron job running on the database server, continues. You want the solution to minimize cost and operations overhead. What should you do?

A.

Use Cloud Scheduler and Cloud Functions to run the daily export.

B.

Create a streaming Datatlow job to export the table.

C.

Set up Cloud Composer, and create a task to export the table daily.

D.

Run the cron job on a Compute Engine instance to continue the export.

Full Access
Question # 9

You are migrating a telehealth care company's on-premises data center to Google Cloud. The migration plan specifies:

PostgreSQL databases must be migrated to a multi-region backup configuration with cross-region replicas to allow restore and failover in multiple scenarios.

MySQL databases handle personally identifiable information (PII) and require data residency compliance at the regional level.

You want to set up the environment with minimal administrative effort. What should you do?

A.

Set up Cloud Logging and Cloud Monitoring with Cloud Functions to send an alert every time a new database instance is created, and manually validate the region.

B.

Set up different organizations for each database type, and apply policy constraints at the organization level.

C.

Set up Pub/Sub to ingest data from Cloud Logging, send an alert every time a new database instance is created, and manually validate the region.

D.

Set up different projects for PostgreSQL and MySQL databases, and apply organizational policy constraints at a project level.

Full Access
Question # 10

You are configuring the networking of a Cloud SQL instance. The only application that connects to this database resides on a Compute Engine VM in the same project as the Cloud SQL instance. The VM and the Cloud SQL instance both use the same VPC network, and both have an external (public) IP address and an internal (private) IP address. You want to improve network security. What should you do?

A.

Disable and remove the internal IP address assignment.

B.

Disable both the external IP address and the internal IP address, and instead rely on Private Google Access.

C.

Specify an authorized network with the CIDR range of the VM.

D.

Disable and remove the external IP address assignment.

Full Access
Question # 11

Your ecommerce website captures user clickstream data to analyze customer traffic patterns in real time and support personalization features on your website. You plan to analyze this data using big data tools. You need a low-latency solution that can store 8 TB of data and can scale to millions of read and write requests per second. What should you do?

A.

Write your data into Bigtable and use Dataproc and the Apache Hbase libraries for analysis.

B.

Deploy a Cloud SQL environment with read replicas for improved performance. Use Datastream to export data to Cloud Storage and analyze with Dataproc and the Cloud Storage connector.

C.

Use Memorystore to handle your low-latency requirements and for real-time analytics.

D.

Stream your data into BigQuery and use Dataproc and the BigQuery Storage API to analyze large volumes of data.

Full Access
Question # 12

You are configuring a brand new PostgreSQL database instance in Cloud SQL. Your application team wants to have an optimal and highly available environment with automatic failover to avoid any unplanned outage. What should you do?

A.

Create one regional Cloud SQL instance with a read replica in another region.

B.

Create one regional Cloud SQL instance in one zone with a standby instance in another zone in the same region.

C.

Create two read-write Cloud SQL instances in two different zones with a standby instance in another region.

D.

Create two read-write Cloud SQL instances in two different regions with a standby instance in another zone.

Full Access
Question # 13

Your DevOps team is using Terraform to deploy applications and Cloud SQL databases. After every new application change is rolled out, the environment is torn down and recreated, and the persistent database layer is lost. You need to prevent the database from being dropped. What should you do?

A.

Set Terraform deletion_protection to true.

B.

Rerun terraform apply.

C.

Create a read replica.

D.

Use point-in-time-recovery (PITR) to recover the database.

Full Access
Question # 14

Your company is migrating their MySQL database to Cloud SQL and cannot afford any planned downtime during the month of December. The company is also concerned with cost, so you need the most cost-effective solution. What should you do?

A.

Open a support ticket in Google Cloud to prevent any maintenance in that MySQL instance during the month of December.

B.

Use Cloud SQL maintenance settings to prevent any maintenance during the month of December.

C.

Create MySQL read replicas in different zones so that, if any downtime occurs, the read replicas will act as the primary instance during the month of December.

D.

Create a MySQL regional instance so that, if any downtime occurs, the standby instance will act as the primary instance during the month of December.

Full Access
Question # 15

Your organization has a busy transactional Cloud SQL for MySQL instance. Your analytics team needs access to the data so they can build monthly sales reports. You need to provide data access to the analytics team without adversely affecting performance. What should you do?

A.

Create a read replica of the database, provide the database IP address, username, and password to the analytics team, and grant read access to required tables to the team.

B.

Create a read replica of the database, enable the cloudsql.iam_authentication flag on the replica, and grant read access to required tables to the analytics team.

C.

Enable the cloudsql.iam_authentication flag on the primary database instance, and grant read access to required tables to the analytics team.

D.

Provide the database IP address, username, and password of the primary database instance to the analytics, team, and grant read access to required tables to the team.

Full Access
Question # 16

Your customer has a global chat application that uses a multi-regional Cloud Spanner instance. The application has recently experienced degraded performance after a new version of the application was launched. Your customer asked you for assistance. During initial troubleshooting, you observed high read latency. What should you do?

A.

Use query parameters to speed up frequently executed queries.

B.

Change the Cloud Spanner configuration from multi-region to single region.

C.

Use SQL statements to analyze SPANNER_SYS.READ_STATS* tables.

D.

Use SQL statements to analyze SPANNER_SYS.QUERY_STATS* tables.

Full Access
Question # 17

You support a consumer inventory application that runs on a multi-region instance of Cloud Spanner. A customer opened a support ticket to complain about slow response times. You notice a Cloud Monitoring alert about high CPU utilization. You want to follow Google-recommended practices to address the CPU performance issue. What should you do first?

A.

Increase the number of processing units.

B.

Modify the database schema, and add additional indexes.

C.

Shard data required by the application into multiple instances.

D.

Decrease the number of processing units.

Full Access
Question # 18

You need to migrate a 1 TB PostgreSQL database from a Compute Engine VM to Cloud SQL for PostgreSQL. You want to ensure that there is minimal downtime during the migration. What should you do?

A.

Export the data from the existing database, and load the data into a new Cloud SQL database.

B.

Use Migrate for Compute Engine to complete the migration.

C.

Use Datastream to complete the migration.

D.

Use Database Migration Service to complete the migration.

Full Access
Question # 19

You are designing a database strategy for a new web application in one region. You need to minimize write latency. What should you do?

A.

Use Cloud SQL with cross-region replicas.

B.

Use high availability (HA) Cloud SQL with multiple zones.

C.

Use zonal Cloud SQL without high availability (HA).

D.

Use Cloud Spanner in a regional configuration.

Full Access
Question # 20

Your company is migrating all legacy applications to Google Cloud. All on-premises applications are using legacy Oracle 12c databases with Oracle Real Application Cluster (RAC) for high availability (HA) and Oracle Data Guard for disaster recovery. You need a solution that requires minimal code changes, provides the same high availability you have today on-premises, and supports a low latency network for migrated legacy applications. What should you do?

A.

Migrate the databases to Cloud Spanner.

B.

Migrate the databases to Cloud SQL, and enable a standby database.

C.

Migrate the databases to Compute Engine using regional persistent disks.

D.

Migrate the databases to Bare Metal Solution for Oracle.

Full Access
Question # 21

You are designing a physician portal app in Node.js. This application will be used in hospitals and clinics that might have intermittent internet connectivity. If a connectivity failure occurs, the app should be able to query the cached data. You need to ensure that the application has scalability, strong consistency, and multi-region replication. What should you do?

A.

Use Firestore and ensure that the PersistenceEnabled option is set to true.

B.

Use Memorystore for Memcached.

C.

Use Pub/Sub to synchronize the changes from the application to Cloud Spanner.

D.

Use Table.read with the exactStaleness option to perform a read of rows in Cloud Spanner.

Full Access
Question # 22

You are designing a database architecture for a global application that stores information about public parks worldwide. The application uses the database for read-only purposes, and a centralized batch job updates the database nightly. You want to select an open source, SQL-compliant database. What should you do?

A.

Use Bigtable with multi-region clusters.

B.

Use Memorystore for Redis with multi-zones within a region.

C.

Use Cloud SQL for PostgreSQL with cross-region replicas.

D.

Use Cloud Spanner with multi-region configuration.

Full Access
Question # 23

You are running a large, highly transactional application on Oracle Real Application Cluster (RAC) that is multi-tenant and uses shared storage. You need a solution that ensures high-performance throughput and a low-latency connection between applications and databases. The solution must also support existing Oracle features and provide ease of migration to Google Cloud. What should you do?

A.

Migrate to Compute Engine.

B.

Migrate to Bare Metal Solution for Oracle.

C.

Migrate to Google Kubernetes Engine (GKE)

D.

Migrate to Google Cloud VMware Engine

Full Access
Question # 24

Your company is developing a new global transactional application that must be ACID-compliant and have 99.999% availability. You are responsible for selecting the appropriate Google Cloud database to serve as a datastore for this new application. What should you do?

A.

Use Firestore.

B.

Use Cloud Spanner.

C.

Use Cloud SQL.

D.

Use Bigtable.

Full Access
Question # 25

You are building an application that allows users to customize their website and mobile experiences. The application will capture user information and preferences. User profiles have a dynamic schema, and users can add or delete information from their profile. You need to ensure that user changes automatically trigger updates to your downstream BigQuery data warehouse. What should you do?

A.

Store your data in Bigtable, and use the user identifier as the key. Use one column family to store user profile data, and use another column family to store user preferences.

B.

Use Cloud SQL, and create different tables for user profile data and user preferences from your recommendations model. Use SQL to join the user profile data and preferences

C.

Use Firestore in Native mode, and store user profile data as a document. Update the user profile with preferences specific to that user and use the user identifier to query.

D.

Use Firestore in Datastore mode, and store user profile data as a document. Update the user profile with preferences specific to that user and use the user identifier to query.

Full Access
Question # 26

Your company uses Cloud Spanner for a mission-critical inventory management system that is globally available. You recently loaded stock keeping unit (SKU) and product catalog data from a company acquisition and observed hot-spots in the Cloud Spanner database. You want to follow Google-recommended schema design practices to avoid performance degradation. What should you do? (Choose two.)

A.

Use an auto-incrementing value as the primary key.

B.

Normalize the data model.

C.

Promote low-cardinality attributes in multi-attribute primary keys.

D.

Promote high-cardinality attributes in multi-attribute primary keys.

E.

Use bit-reverse sequential value as the primary key.

Full Access
Question # 27

You are configuring a brand new Cloud SQL for PostgreSQL database instance in Google Cloud. Your application team wants you to deploy one primary instance, one standby instance, and one read replica instance. You need to ensure that you are following Google-recommended practices for high availability. What should you do?

A.

Configure the primary instance in zone A, the standby instance in zone C, and the read replica in zone B, all in the same region.

B.

Configure the primary and standby instances in zone A and the read replica in zone B, all in the same region.

C.

Configure the primary instance in one region, the standby instance in a second region, and the read replica in a third region.

D.

Configure the primary, standby, and read replica instances in zone A, all in the same region.

Full Access
Question # 28

You are migrating an on-premises application to Google Cloud. The application requires a high availability (HA) PostgreSQL database to support business-critical functions. Your company's disaster recovery strategy requires a recovery time objective (RTO) and recovery point objective (RPO) within 30 minutes of failure. You plan to use a Google Cloud managed service. What should you do to maximize uptime for your application?

A.

Deploy Cloud SQL for PostgreSQL in a regional configuration. Create a read replica in a different zone in the same region and a read replica in another region for disaster recovery.

B.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Take periodic backups, and use this backup to restore to a new Cloud SQL for PostgreSQL instance in another region during a disaster recovery event.

C.

Deploy Cloud SQL for PostgreSQL in a regional configuration with HA enabled. Create a cross-region read replica, and promote the read replica as the primary node for disaster recovery.

D.

Migrate the PostgreSQL database to multi-regional Cloud Spanner so that a single region outage will not affect your application. Update the schema to support Cloud Spanner data types, and refactor the application.

Full Access
Question # 29

You are migrating an on-premises application to Compute Engine and Cloud SQL. The application VMs will live in their own project, separate from the Cloud SQL instances which have their own project. What should you do to configure the networks?

A.

Create a new VPC network in each project, and use VPC Network Peering to connect the two together.

B.

Create a Shared VPC that both the application VMs and Cloud SQL instances will use.

C.

Use the default networks, and leverage Cloud VPN to connect the two together.

D.

Place both the application VMs and the Cloud SQL instances in the default network of each project.

Full Access
Question # 30

You are designing a highly available (HA) Cloud SQL for PostgreSQL instance that will be used by 100 databases. Each database contains 80 tables that were migrated from your on-premises environment to Google Cloud. The applications that use these databases are located in multiple regions in the US, and you need to ensure that read and write operations have low latency. What should you do?

A.

Deploy 2 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-east1 and us-west1.

B.

Deploy 2 Cloud SQL instances in the us-central1 region, and create read replicas in us-east1 and us-west1.

C.

Deploy 4 Cloud SQL instances in the us-central1 region with HA enabled, and create read replicas in us-central1, us-east1, and us-west1.

D.

Deploy 4 Cloud SQL instances in the us-central1 region, and create read replicas in us-central1, us-east1 and us-west1.

Full Access
Question # 31

Your organization is migrating 50 TB Oracle databases to Bare Metal Solution for Oracle. Database backups must be available for quick restore. You also need to have backups available for 5 years. You need to design a cost-effective architecture that meets a recovery time objective (RTO) of 2 hours and recovery point objective (RPO) of 15 minutes. What should you do?

A.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on all flash storage.

Keep backups older than one day stored in Actifio OnVault storage.

B.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on standard storage.

Keep backups older than one day stored in Actifio OnVault storage.

C.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on standard storage.

Use the Oracle Recovery Manager (RMAN) backup utility to move backups older than one day to a Coldline Storage bucket.

D.

Create the database on a Bare Metal Solution server with the database running on flash storage.

Keep a local backup copy on all flash storage.

Use the Oracle Recovery Manager (RMAN) backup utility to move backups older than one day to an Archive Storage bucket.

Full Access
Question # 32

You want to migrate an on-premises 100 TB Microsoft SQL Server database to Google Cloud over a 1 Gbps network link. You have 48 hours allowed downtime to migrate this database. What should you do? (Choose two.)

A.

Use a change data capture (CDC) migration strategy.

B.

Move the physical database servers from on-premises to Google Cloud.

C.

Keep the network bandwidth at 1 Gbps, and then perform an offline data migration.

D.

Increase the network bandwidth to 2 Gbps, and then perform an offline data migration.

E.

Increase the network bandwidth to 10 Gbps, and then perform an offline data migration.

Full Access
Question # 33

Your organization has a ticketing system that needs an online marketing analytics and reporting application. You need to select a relational database that can manage hundreds of terabytes of data to support this new application. Which database should you use?

A.

Cloud SQL

B.

BigQuery

C.

Cloud Spanner

D.

Bigtable

Full Access
Question # 34

Your company is shutting down their on-premises data center and migrating their Oracle databases using Oracle Real Application Clusters (RAC) to Google Cloud. You want minimal to no changes to the applications during the database migration. What should you do?

A.

Migrate the Oracle databases to Cloud Spanner.

B.

Migrate the Oracle databases to Compute Engine.

C.

Migrate the Oracle databases to Cloud SQL.

D.

Migrate the Oracle databases to Bare Metal Solution for Oracle.

Full Access
Question # 35

You are writing an application that will run on Cloud Run and require a database running in the Cloud SQL managed service. You want to secure this instance so that it only receives connections from applications running in your VPC environment in Google Cloud. What should you do?

A.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Use Cloud SQL Auth proxy to connect to the instance.

B.

1. Create your instance with a specified external (public) IP address.

2. Choose the VPC and create firewall rules to allow only connections from Cloud Run into your instance.

3. Connect to the instance using a connection pool to best manage connections to the instance.

C.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Use Cloud SQL Auth proxy to connect to the instance.

D.

1. Create your instance with a specified internal (private) IP address.

2. Choose the VPC with private service connection configured.

3. Configure the Serverless VPC Access connector in the same VPC network as your Cloud SQL instance.

4. Connect to the instance using a connection pool to best manage connections to the instance.

Full Access
Question # 36

Your organization is running a Firestore-backed Firebase app that serves the same top ten news stories on a daily basis to a large global audience. You want to optimize content delivery while decreasing cost and latency. What should you do?

A.

Enable serializable isolation in the Firebase app.

B.

Deploy a US multi-region Firestore location.

C.

Build a Firestore bundle, and deploy bundles to Cloud CDN.

D.

Create a Firestore index on the news story date.

Full Access
Question # 37

You released a popular mobile game and are using a 50 TB Cloud Spanner instance to store game data in a PITR-enabled production environment. When you analyzed the game statistics, you realized that some players are exploiting a loophole to gather more points to get on the leaderboard. Another DBA accidentally ran an emergency bugfix script that corrupted some of the data in the production environment. You need to determine the extent of the data corruption and restore the production environment. What should you do? (Choose two.)

A.

If the corruption is significant, use backup and restore, and specify a recovery timestamp.

B.

If the corruption is significant, perform a stale read and specify a recovery timestamp. Write the results back.

C.

If the corruption is significant, use import and export.

D.

If the corruption is insignificant, use backup and restore, and specify a recovery timestamp.

E.

If the corruption is insignificant, perform a stale read and specify a recovery timestamp. Write the results back.

Full Access
Question # 38

Your organization operates in a highly regulated industry. Separation of concerns (SoC) and security principle of least privilege (PoLP) are critical. The operations team consists of:

Person A is a database administrator.

Person B is an analyst who generates metric reports.

Application C is responsible for automatic backups.

You need to assign roles to team members for Cloud Spanner. Which roles should you assign?

A.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseReader for Person B

roles/spanner.backupWriter for Application C

B.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseReader for Person B

roles/spanner.backupAdmin for Application C

C.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseUser for Person B

roles/spanner databaseReader for Application C

D.

roles/spanner.databaseAdmin for Person A

roles/spanner.databaseUser for Person B

roles/spanner.backupWriter for Application C

Full Access
Question # 39

Your organization is currently updating an existing corporate application that is running in another public cloud to access managed database services in Google Cloud. The application will remain in the other public cloud while the database is migrated to Google Cloud. You want to follow Google-recommended practices for authentication. You need to minimize user disruption during the migration. What should you do?

A.

Use workload identity federation to impersonate a service account.

B.

Ask existing users to set their Google password to match their corporate password.

C.

Migrate the application to Google Cloud, and use Identity and Access Management (IAM).

D.

Use Google Workspace Password Sync to replicate passwords into Google Cloud.

Full Access