Which environment should you train your model on?

Your team is building a convolutional neural network (CNN)-based architecture from scratch. The preliminary experiments running on your on-premises CPU-only infrastructure were encouraging, but have slow convergence. You have been asked to speed up model training to reduce time-to-market. You want to experiment with virtual machines (VMs) on Google Cloud to leverage more powerful hardware. Your code does not include any manual device placement and has not been wrapped in Estimator model-level abstraction.

Which environment should you train your model on?
A . AVM on Compute Engine and 1 TPU with all dependencies installed manually.
B . AVM on Compute Engine and 8 GPUs with all dependencies installed manually.
C . A Deep Learning VM with an n1-standard-2 machine and 1 GPU with all libraries pre-installed.
D . A Deep Learning VM with more powerful CPU e2-highcpu-16 machines with all libraries pre-installed.

Answer: C

Explanation:

https://cloud.google.com/deep-learning-vm/docs/cli#creating_an_instance_with_one_or_more_gpus

https://cloud.google.com/deep-learning-vm/docs/introduction#pre-installed_packages

"speed up model training" will make us biased towards GPU,TPU options by options eliminations we may need to stay away of any manual installations , so using preconfigered deep learning will speed up time to market

Subscribe
Notify of
guest
0 Comments
Inline Feedbacks
View all comments