Special Summer Sale Discount Flat 70% Offer - Ends in 0d 00h 00m 00s - Coupon code: 70diswrap

Google Associate-Data-Practitioner Dumps

Google Cloud Associate Data Practitioner (ADP Exam) Questions and Answers

Question 1

Your team needs to analyze large datasets stored in BigQuery to identify trends in user behavior. The analysis will involve complex statistical calculations, Python packages, and visualizations. You need to recommend a managed collaborative environment to develop and share the analysis. What should you recommend?

Options:

A.

Create a Colab Enterprise notebook and connect the notebook to BigQuery. Share the notebook with your team. Analyze the data and generate visualizations in Colab Enterprise.

B.

Create a statistical model by using BigQuery ML. Share the query with your team. Analyze the data and generate visualizations in Looker Studio.

C.

Create a Looker Studio dashboard and connect the dashboard to BigQuery. Share the dashboard with your team. Analyze the data and generate visualizations in Looker Studio.

D.

Connect Google Sheets to BigQuery by using Connected Sheets. Share the Google Sheet with your team. Analyze the data and generate visualizations in Gooqle Sheets.

Question 2

You manage a BigQuery table that is used for critical end-of-month reports. The table is updated weekly with new sales data. You want to prevent data loss and reporting issues if the table is accidentally deleted. What should you do?

Options:

A.

Configure the time travel duration on the table to be exactly seven days. On deletion, re-create the deleted table solely from the time travel data.

B.

Schedule the creation of a new snapshot of the table once a week. On deletion, re-create the deleted table using the snapshot and time travel data.

C.

Create a clone of the table. On deletion, re-create the deleted table by copying the content of the clone.

D.

Create a view of the table. On deletion, re-create the deleted table from the view and time travel data.

Question 3

You need to create a new data pipeline. You want a serverless solution that meets the following requirements:

• Data is streamed from Pub/Sub and is processed in real-time.

• Data is transformed before being stored.

• Data is stored in a location that will allow it to be analyzed with SQL using Looker.

Which Google Cloud services should you recommend for the pipeline?

Options:

A.

1. Dataproc Serverless

2. Bigtable

B.

1. Cloud Composer

2. Cloud SQL for MySQL

C.

1. BigQuery

2. Analytics Hub

D.

1. Dataflow

2. BigQuery

Question 4

Your organization has highly sensitive data that gets updated once a day and is stored across multiple datasets in BigQuery. You need to provide a new data analyst access to query specific data in BigQuery while preventing access to sensitive data. What should you do?

Options:

A.

Grant the data analyst the BigQuery Job User IAM role in the Google Cloud project.

B.

Create a materialized view with the limited data in a new dataset. Grant the data analyst BigQuery Data Viewer IAM role in the dataset and the BigQuery Job User IAM role in the Google Cloud project.

C.

Create a new Google Cloud project, and copy the limited data into a BigQuery table. Grant the data analyst the BigQuery Data Owner IAM role in the new Google Cloud project.

D.

Grant the data analyst the BigQuery Data Viewer IAM role in the Google Cloud project.

Question 5

You need to create a weekly aggregated sales report based on a large volume of data. You want to use Python to design an efficient process for generating this report. What should you do?

Options:

A.

Create a Cloud Run function that uses NumPy. Use Cloud Scheduler to schedule the function to run once a week.

B.

Create a Colab Enterprise notebook and use the bigframes.pandas library. Schedule the notebook to execute once a week.

C.

Create a Cloud Data Fusion and Wrangler flow. Schedule the flow to run once a week.

D.

Create a Dataflow directed acyclic graph (DAG) coded in Python. Use Cloud Scheduler to schedule the code to run once a week.

Question 6

You need to create a data pipeline that streams event information from applications in multiple Google Cloud regions into BigQuery for near real-time analysis. The data requires transformation before loading. You want to create the pipeline using a visual interface. What should you do?

Options:

A.

Push event information to a Pub/Sub topic. Create a Dataflow job using the Dataflow job builder.

B.

Push event information to a Pub/Sub topic. Create a Cloud Run function to subscribe to the Pub/Sub topic, apply transformations, and insert the data into BigQuery.

C.

Push event information to a Pub/Sub topic. Create a BigQuery subscription in Pub/Sub.

D.

Push event information to Cloud Storage, and create an external table in BigQuery. Create a BigQuery scheduled job that executes once each day to apply transformations.

Question 7

Your organization stores highly personal data in BigQuery and needs to comply with strict data privacy regulations. You need to ensure that sensitive data values are rendered unreadable whenever an employee leaves the organization. What should you do?

Options:

A.

Use AEAD functions and delete keys when employees leave the organization.

B.

Use dynamic data masking and revoke viewer permissions when employees leave the organization.

C.

Use customer-managed encryption keys (CMEK) and delete keys when employees leave the organization.

D.

Use column-level access controls with policy tags and revoke viewer permissions when employees leave the organization.

Question 8

You manage a web application that stores data in a Cloud SQL database. You need to improve the read performance of the application by offloading read traffic from the primary database instance. You want to implement a solution that minimizes effort and cost. What should you do?

Options:

A.

Use Cloud CDN to cache frequently accessed data.

B.

Store frequently accessed data in a Memorystore instance.

C.

Migrate the database to a larger Cloud SQL instance.

D.

Enable automatic backups, and create a read replica of the Cloud SQL instance.

Question 9

You are a database administrator managing sales transaction data by region stored in a BigQuery table. You need to ensure that each sales representative can only see the transactions in their region. What should you do?

Options:

A.

Add a policy tag in BigQuery.

B.

Create a row-level access policy.

C.

Create a data masking rule.

D.

Grant the appropriate 1AM permissions on the dataset.

Question 10

You have created a LookML model and dashboard that shows daily sales metrics for five regional managers to use. You want to ensure that the regional managers can only see sales metrics specific to their region. You need an easy-to-implement solution. What should you do?

Options:

A.

Create asales_regionuser attribute, and assign each manager’s region as the value of their user attribute. Add anaccess_filterExplore filter on theregion_namedimension by using thesales_regionuser attribute.

B.

Create five different Explores with thesql_always_filterExplore filter applied on theregion_namedimension. Set eachregion_namevalue to the corresponding region for each manager.

C.

Create separate Looker dashboards for each regional manager. Set the default dashboard filter to the corresponding region for each manager.

D.

Create separate Looker instances for each regional manager. Copy the LookML model and dashboard to each instance. Provision viewer access to the corresponding manager.

Question 11

Your company currently uses an on-premises network file system (NFS) and is migrating data to Google Cloud. You want to be able to control how much bandwidth is used by the data migration while capturing detailed reporting on the migration status. What should you do?

Options:

A.

Use a Transfer Appliance.

B.

Use Cloud Storage FUSE.

C.

Use Storage Transfer Service.

D.

Use gcloud storage commands.

Question 12

You are constructing a data pipeline to process sensitive customer data stored in a Cloud Storage bucket. You need to ensure that this data remains accessible, even in the event of a single-zone outage. What should you do?

Options:

A.

Set up a Cloud CDN in front of the bucket.

B.

Enable Object Versioning on the bucket.

C.

Store the data in a multi-region bucket.

D.

Store the data in Nearline storaqe.

Question 13

Your organization has a BigQuery dataset that contains sensitive employee information such as salaries and performance reviews. The payroll specialist in the HR department needs to have continuous access to aggregated performance data, but they do not need continuous access to other sensitive data. You need to grant the payroll specialist access to the performance data without granting them access to the entire dataset using the simplest and most secure approach. What should you do?

Options:

A.

Use authorized views to share query results with the payroll specialist.

B.

Create row-level and column-level permissions and policies on the table that contains performance data in the dataset. Provide the payroll specialist with the appropriate permission set.

C.

Create a table with the aggregated performance data. Use table-level permissions to grant access to the payroll specialist.

D.

Create a SQL query with the aggregated performance data. Export the results to an Avro file in a Cloud Storage bucket. Share the bucket with the payroll specialist.

Question 14

You work for a retail company that collects customer data from various sources:

    Online transactions: Stored in a MySQL database

    Customer feedback: Stored as text files on a company server

    Social media activity: Streamed in real-time from social media platformsYou need to design a data pipeline to extract and load the data into the appropriate Google Cloud storage system(s) for further analysis and ML model training. What should you do?

Options:

A.

Copy the online transactions data into Cloud SQL for MySQL. Import the customer feedback into BigQuery. Stream the social media activity into Cloud Storage.

B.

Extract and load the online transactions data into BigQuery. Load the customer feedback data into Cloud Storage. Stream the social media activity by using Pub/Sub and Dataflow, and store the data in BigQuery.

C.

Extract and load the online transactions data, customer feedback data, and social media activity into Cloud Storage.

D.

Extract and load the online transactions data into Bigtable. Import the customer feedback data into Cloud Storage. Store the social media activity in Cloud SQL for MySQL.

Question 15

You need to design a data pipeline that ingests data from CSV, Avro, and Parquet files into Cloud Storage. The data includes raw user input. You need to remove all malicious SQL injections before storing the data in BigQuery. Which data manipulation methodology should you choose?

Options:

A.

EL

B.

ELT

C.

ETL

D.

ETLT

Question 16

Your team wants to create a monthly report to analyze inventory data that is updated daily. You need to aggregate the inventory counts by using only the most recent month of data, and save the results to be used in a Looker Studio dashboard. What should you do?

Options:

A.

Create a materialized view in BigQuery that uses the SUM( ) function and the DATE_SUB( ) function.

B.

Create a saved query in the BigQuery console that uses the SUM( ) function and the DATE_SUB( ) function. Re-run the saved query every month, and save the results to a BigQuery table.

C.

Create a BigQuery table that uses the SUM( ) function and the _PARTITIONDATE filter.

D.

Create a BigQuery table that uses the SUM( ) function and the DATE_DIFF( ) function.

Question 17

Your organization needs to store historical customer order data. The data will only be accessed once a month for analysis and must be readily available within a few seconds when it is accessed. You need to choose a storage class that minimizes storage costs while ensuring that the data can be retrieved quickly. What should you do?

Options:

A.

Store the data in Cloud Storaqe usinq Nearline storaqe.

B.

Store the data in Cloud Storaqe usinq Coldline storaqe.

C.

Store the data in Cloud Storage using Standard storage.

D.

Store the data in Cloud Storage using Archive storage.

Question 18

You are designing a BigQuery data warehouse with a team of experienced SQL developers. You need to recommend a cost-effective, fully-managed, serverless solution to build ELT processes with SQL pipelines. Your solution must include source code control, environment parameterization, and data quality checks. What should you do?

Options:

A.

Use Cloud Data Fusion to visually design and manage the pipelines.

B.

Use Dataform to build, orchestrate, and monitor the pipelines.

C.

Use Dataproc to run MapReduce jobs for distributed data processing.

D.

Use Cloud Composer to orchestrate and run data workflows.

Question 19

Your organization has decided to move their on-premises Apache Spark-based workload to Google Cloud. You want to be able to manage the code without needing to provision and manage your own cluster. What should you do?

Options:

A.

Migrate the Spark jobs to Dataproc Serverless.

B.

Configure a Google Kubernetes Engine cluster with Spark operators, and deploy the Spark jobs.

C.

Migrate the Spark jobs to Dataproc on Google Kubernetes Engine.

D.

Migrate the Spark jobs to Dataproc on Compute Engine.

Question 20

You want to process and load a daily sales CSV file stored in Cloud Storage into BigQuery for downstream reporting. You need to quickly build a scalable data pipeline that transforms the data while providing insights into data quality issues. What should you do?

Options:

A.

Create a batch pipeline in Cloud Data Fusion by using a Cloud Storage source and a BigQuery sink.

B.

Load the CSV file as a table in BigQuery, and use scheduled queries to run SQL transformation scripts.

C.

Load the CSV file as a table in BigQuery. Create a batch pipeline in Cloud Data Fusion by using a BigQuery source and sink.

D.

Create a batch pipeline in Dataflow by using the Cloud Storage CSV file to BigQuery batch template.

Question 21

You need to create a data pipeline for a new application. Your application will stream data that needs to be enriched and cleaned. Eventually, the data will be used to train machine learning models. You need to determine the appropriate data manipulation methodology and which Google Cloud services to use in this pipeline. What should you choose?

Options:

A.

ETL; Dataflow -> BigQuery

B.

ETL; Cloud Data Fusion -> Cloud Storage

C.

ELT; Cloud Storage -> Bigtable

D.

ELT; Cloud SQL -> Analytics Hub

Question 22

Your organization has several datasets in their data warehouse in BigQuery. Several analyst teams in different departments use the datasets to run queries. Your organization is concerned about the variability of their monthly BigQuery costs. You need to identify a solution that creates a fixed budget for costs associated with the queries run by each department. What should you do?

Options:

A.

Create a custom quota for each analyst in BigQuery.

B.

Create a single reservation by using BigQuery editions. Assign all analysts to the reservation.

C.

Assign each analyst to a separate project associated with their department. Create a single reservation by using BigQuery editions. Assign all projects to the reservation.

D.

Assign each analyst to a separate project associated with their department. Create a single reservation for each department by using BigQuery editions. Create assignments for each project in the appropriate reservation.

Question 23

You are a data analyst working with sensitive customer data in BigQuery. You need to ensure that only authorized personnel within your organization can query this data, while following the principle of least privilege. What should you do?

Options:

A.

Enable access control by using IAM roles.

B.

Update dataset privileges by using the SQL GRANT statement.

C.

Export the data to Cloud Storage, and use signed URLs to authorize access.

D.

Encrypt the data by using customer-managed encryption keys (CMEK).

Question 24

Your organization’s business analysts require near real-time access to streaming data. However, they are reporting that their dashboard queries are loading slowly. After investigating BigQuery query performance, you discover the slow dashboard queries perform several joins and aggregations.

You need to improve the dashboard loading time and ensure that the dashboard data is as up-to-date as possible. What should you do?

Options:

A.

Disable BiqQuery query result caching.

B.

Modify the schema to use parameterized data types.

C.

Create a scheduled query to calculate and store intermediate results.

D.

Create materialized views.

Question 25

You want to build a model to predict the likelihood of a customer clicking on an online advertisement. You have historical data in BigQuery that includes features such as user demographics, ad placement,and previous click behavior. After training the model, you want to generate predictions on new data. Which model type should you use in BigQuery ML?

Options:

A.

Linear regression

B.

Matrix factorization

C.

Logistic regression

D.

K-means clustering

Question 26

You have millions of customer feedback records stored in BigQuery. You want to summarize the data by using the large language model (LLM) Gemini. You need to plan and execute this analysis using the most efficient approach. What should you do?

Options:

A.

Query the BigQuery table from within a Python notebook, use the Gemini API to summarize the data within the notebook, and store the summaries in BigQuery.

B.

Use a BigQuery ML model to pre-process the text data, export the results to Cloud Storage, and use the Gemini API to summarize the pre- processed data.

C.

Create a BigQuery Cloud resource connection to a remote model in Vertex Al, and use Gemini to summarize the data.

D.

Export the raw BigQuery data to a CSV file, upload it to Cloud Storage, and use the Gemini API to summarize the data.

Page: 1 / 11
Total 106 questions