New Year Sale 2026! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Google Associate Cloud Engineer Exam - Topic 2 Question 111 Discussion

Actual exam question for Google's Associate Cloud Engineer exam
Question #: 111
Topic #: 2
[All Associate Cloud Engineer Questions]

You have a large 5-TB AVRO file stored in a Cloud Storage bucket. Your analysts are proficient only in SQL and need access to the data stored in this file. You want to find a cost-effective way to complete their request as soon as possible. What should you do?

Show Suggested Answer Hide Answer
Suggested Answer: C

https://cloud.google.com/bigquery/external-data-sources

An external data source is a data source that you can query directly from BigQuery, even though the data is not stored in BigQuery storage.

BigQuery supports the following external data sources:

Amazon S3

Azure Storage

Cloud Bigtable

Cloud Spanner

Cloud SQL

Cloud Storage

Drive


Contribute your Thoughts:

0/2000 characters
Madelyn
2 months ago
Wait, why not just use external tables? C seems easier.
upvoted 0 times
...
Herman
2 months ago
Definitely B! Quick and cost-effective.
upvoted 0 times
...
Carmela
3 months ago
Surprised there's no mention of using Dataflow here!
upvoted 0 times
...
Elvera
3 months ago
I thought Hadoop was outdated for this kind of task?
upvoted 0 times
...
Francine
3 months ago
BigQuery is super efficient for large datasets!
upvoted 0 times
...
Barney
3 months ago
I'm leaning towards option B, but I worry about the costs of loading such a large file into BigQuery just to drop it afterward.
upvoted 0 times
...
Yolande
4 months ago
I feel like we practiced a similar question where using external tables in BigQuery was the right choice, but I can't recall the exact details.
upvoted 0 times
...
Freeman
4 months ago
I think option C sounds familiar because it allows querying directly from Cloud Storage without needing to load the data, which seems efficient.
upvoted 0 times
...
Lavonna
4 months ago
I remember we discussed using BigQuery for large datasets, but I'm not sure if creating a table and dropping it later is the best approach for cost-effectiveness.
upvoted 0 times
...
Jennifer
4 months ago
Option D with Hadoop seems like overkill for this use case. The data is already in a Cloud Storage bucket, so I think the best approach is to use BigQuery, either through external tables or by loading the data directly.
upvoted 0 times
...
Howard
4 months ago
I'm leaning towards option B. Creating a temporary BigQuery table and loading the data there might be a bit more work, but it could provide better performance for the analysts' SQL queries.
upvoted 0 times
...
Reta
4 months ago
Option C seems like the most straightforward solution. Creating external tables in BigQuery that point to the Cloud Storage bucket should allow the analysts to run SQL queries without having to load the entire 5-TB file.
upvoted 0 times
...
Paulene
5 months ago
This looks like a tricky one. I'm not sure if I should go with option B or C. I'll need to think through the pros and cons of each approach.
upvoted 0 times
...
Lashaun
5 months ago
I disagree, I believe option C is better. It allows direct querying of data in Cloud Storage.
upvoted 0 times
...
Katina
6 months ago
I think option B is the best choice. It's cost-effective and efficient.
upvoted 0 times
...

Save Cancel