Welcome to ExamTopics
ExamTopics Logo
- Expert Verified, Online, Free.
exam questions

Exam Professional Cloud Architect All Questions

View all questions & answers for the Professional Cloud Architect exam

Exam Professional Cloud Architect topic 8 question 4 discussion

Actual exam question from Google's Professional Cloud Architect
Question #: 4
Topic #: 8
[All Professional Cloud Architect Questions]

TerramEarth plans to connect all 20 million vehicles in the field to the cloud. This increases the volume to 20 million 600 byte records a second for 40 TB an hour.
How should you design the data ingestion?

  • A. Vehicles write data directly to GCS
  • B. Vehicles write data directly to Google Cloud Pub/Sub
  • C. Vehicles stream data directly to Google BigQuery
  • D. Vehicles continue to write data using the existing system (FTP)
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
jcmoranp
Highly Voted 5 years, 1 month ago
It's Pub/Sub, too much data streaming for Bigquery...
upvoted 41 times
alexspam88
3 years, 5 months ago
Too much for pubsub either https://cloud.google.com/pubsub/quotas
upvoted 4 times
Bill831231
3 years, 1 month ago
thanks for sharing the link, but seems pub/sub can handle more streaming data than bigquery. pub/sub 120,000,000 kB per minute (2 GB/s) in large regions, bigquery is 1GB/s
upvoted 7 times
...
...
...
JoeShmoe
Highly Voted 5 years ago
Its B, it exceeds the streaming limit for BQ
upvoted 20 times
...
VegasDegenerate
Most Recent 4 months, 3 weeks ago
Has to be pub-sub, you have remote vehicles and need to guarantee message delivery.
upvoted 1 times
...
the1dv
7 months, 3 weeks ago
Wow its almost like GCP shouldnt have offloaded their IoT Core product - you cant "Write direct to PubSub". Its the correct answer but its overly simplified Writing directly to GCS will cost a fortune to retrieve in GET requests etc
upvoted 2 times
...
Vesta1807
11 months ago
Selected Answer: C
Streamed data is available for real-time analysis within a few seconds of the first streaming insertion into a table. Instead of using a job to load data into BigQuery, you can choose to stream your data into BigQuery one record at a time by using the tabledata().insertAll() method. This approach enables querying data without the delay of running a load job. References: https://cloud.google.com/bigquery/streaming-data-into-bigquery
upvoted 1 times
...
MahAli
11 months, 2 weeks ago
Selected Answer: A
They are sending files through FTP why everyone is missing this point? The max message size in pub sub is 10MB as I remember, I would keep the files solution and try to roll out updates to direct the upload to GCS
upvoted 5 times
...
BiddlyBdoyng
1 year, 5 months ago
So many people pointing out this breaks the BigQuery quota limit but very few pointing out it also breaks the Pub/Sub quote limit.......... So the answer is either not bound by the quota limit (in which case why not BigQuery) both are wrong and we stick with FTP
upvoted 1 times
...
kapara
1 year, 6 months ago
Selected Answer: B
it's B
upvoted 1 times
...
nunopires2001
1 year, 10 months ago
I know it's B, however the sensors are probably legacy systems, that can not communicate to a pub/sub queue. Ignoring how huge is to change or adapta 20 million devices is a mistake.
upvoted 3 times
...
omermahgoub
1 year, 11 months ago
To handle the volume of data that TerramEarth plans to ingest, it is recommended to use a scalable and reliable data ingestion solution such as Google Cloud Pub/Sub. With Cloud Pub/Sub, the vehicles can stream data directly to the service, which can handle the high volume of data and provide a buffer to absorb sudden spikes in traffic. The data can then be processed and stored in a data warehouse such as BigQuery for analysis. Option A (writing data directly to GCS) may not be suitable for handling high volumes of data in real-time and may result in data loss if the volume exceeds the capacity of GCS. Option C (streaming data directly to BigQuery) may not be suitable for handling high volumes of data in real-time as it may result in data loss or ingestion delays. Option D (continuing to write data using the existing system) may not be suitable as the current system may not be able to handle the increased volume of data and may result in data loss or ingestion delays.
upvoted 10 times
sank8
1 year, 11 months ago
correct. thanks for the explanation
upvoted 1 times
...
...
surajkrishnamurthy
1 year, 11 months ago
Selected Answer: B
B is the correct answer
upvoted 1 times
...
megumin
2 years ago
Selected Answer: B
ok for B
upvoted 1 times
...
Mahmoud_E
2 years, 1 month ago
Selected Answer: B
B is the correct answer, this similar question was in google simple questions
upvoted 1 times
...
AzureDP900
2 years, 4 months ago
B is right!
upvoted 2 times
...
cdcollector
2 years, 5 months ago
Should be A - see next question on 80% cellular connectivity and Avro format files streamed directly to GCS
upvoted 2 times
...
amxexam
2 years, 6 months ago
Selected Answer: B
We need to buffer, the default limit of BigQuery is 100 API calls per second, till now this cannot be changed. Hence we should ease using Pub/Sub so B.
upvoted 2 times
...
[Removed]
2 years, 7 months ago
Selected Answer: B
You can request limit increases to use BQ streaming for this load, but why pay to store data before ETL?
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...