exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 212 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 212
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You are pre-training a large language model on Google Cloud. This model includes custom TensorFlow operations in the training loop. Model training will use a large batch size, and you expect training to take several weeks. You need to configure a training architecture that minimizes both training time and compute costs. What should you do?

  • A. Implement 8 workers of a2-megagpu-16g machines by using tf.distribute.MultiWorkerMirroredStrategy.
  • B. Implement a TPU Pod slice with -accelerator-type=v4-l28 by using tf.distribute.TPUStrategy.
  • C. Implement 16 workers of c2d-highcpu-32 machines by using tf.distribute.MirroredStrategy.
  • D. Implement 16 workers of a2-highgpu-8g machines by using tf.distribute.MultiWorkerMirroredStrategy.
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
pikachu007
Highly Voted 11 months, 2 weeks ago
Selected Answer: B
TPU Advantages: Highly Specialized: TPUs (Tensor Processing Units) are custom-designed hardware accelerators specifically optimized for machine learning workloads, particularly those involving large batch sizes and matrix-heavy computations, common in large language models. Exceptional Performance: TPUs can significantly outperform CPUs and GPUs in terms of speed and efficiency for these types of tasks. Cost-Effective: While TPUs might have a higher hourly cost, their exceptional performance often leads to lower overall costs due to faster training times and reduced resource usage. TPU Pod Slice: Scalability: TPU Pod slices allow you to distribute training across multiple TPUv4 chips for even greater performance and scalability. Custom Operations: The tf.distribute.TPUStrategy ensures compatibility with custom TensorFlow operations,
upvoted 6 times
...
Omi_04040
Most Recent 1 week, 5 days ago
Selected Answer: A
The question says "model includes custom TensorFlow operations in the training loop", this is not supported by TPU. Hence A
upvoted 2 times
...
Pau1234
1 week, 5 days ago
Selected Answer: D
TPUs are not suitable since we are talking about customer operations. Then between A and D. I'd go with D, because it is more cost effective than A. 16g will be more expensive.
upvoted 1 times
...
9fbd29a
3 weeks, 5 days ago
Selected Answer: A
TPUs not recommended for custom operations
upvoted 1 times
...
DaleR
4 weeks, 1 day ago
B is wrong:
upvoted 1 times
...
f084277
1 month, 1 week ago
All the people voting B are wrong. TPUs cannot be used with TF custom operations
upvoted 2 times
...
baimus
3 months, 1 week ago
Selected Answer: A
This could be A or D, because they both will perform will with custom Tensorflow operations. A is likely to be better with large batch sizes, which require bigger GPUs, so I went A.
upvoted 1 times
...
AK2020
4 months, 3 weeks ago
Selected Answer: A
B is not correct as TPUs not suitable for TensorFlow custom operations and C doesn't make any sense. A or D?. I would go with A
upvoted 3 times
...
info_appsatori
6 months, 1 week ago
Should be A or D. TPU is ok, but TPUs not suitable for TensorFlow custom operations.
upvoted 2 times
...
ccb23cc
6 months, 1 week ago
Selected Answer: A
B. TPU Acceleration: the question says that uses Tensorflow custom operations in the main loop and Google documentation literatelly says about TPU use: "Models with no custom TensorFlow/PyTorch/JAX operations inside the main training loop" C. High-CPU Machines: Make no sense because tell you to use a cpu (which does not help us in this case) So the correct answer is between A and D. However the question says that they are planning to use a large batch size so we need RAM. Therefore we should take the one with more. Correct answer: Option A
upvoted 3 times
...
fitri001
8 months, 1 week ago
Selected Answer: B
TPU Acceleration: TPUs are specifically designed for machine learning workloads and offer significant speedups compared to GPUs or CPUs, especially for large models like yours. Utilizing a TPU Pod slice provides access to a collection of interconnected TPUs for efficient parallel training. tf.distribute.TPUStrategy: This strategy is specifically designed to work with TPUs in TensorFlow. It handles data distribution, model replication, and gradient aggregation across the TPU cores, enabling efficient training with custom TensorFlow operations.
upvoted 1 times
fitri001
8 months, 1 week ago
why not the others? A. MultiWorkerMirroredStrategy with GPUs: While GPUs offer some acceleration, TPUs are generally better suited for large language model pre-training due to their architectural optimizations. Additionally, managing 8 workers across separate machines can introduce communication overhead compared to a tightly coupled TPU Pod. C. MirroredStrategy with High-CPU Machines: CPU-based training would be significantly slower than TPUs or even GPUs for a large language model. While the high CPU count might seem beneficial for custom operations, the overall training speed would still be limited. D. MultiWorkerMirroredStrategy with Multiple High-GPU Machines: Similar to option A, using multiple high-GPU machines with this strategy would incur communication overhead and potentially be less cost-effective compared to a single TPU Pod slice.
upvoted 2 times
...
...
BlehMaks
11 months, 1 week ago
Selected Answer: B
It should be TPU but i'm a bit concerned about this point from Google documentation: Models with no custom TensorFlow/PyTorch/JAX operations inside the main training loop https://cloud.google.com/tpu/docs/intro-to-tpu#TPU
upvoted 2 times
...
b1a8fae
11 months, 1 week ago
Selected Answer: B
B. NGL quite lost on this one but if the training set is big enough to span over several weeks I would go with the most powerful resource (TPUs) but I might be completely wrong.
upvoted 3 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago