Deal of The Day! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Google Exam Professional Machine Learning Engineer Topic 5 Question 75 Discussion

Actual exam question for Google's Professional Machine Learning Engineer exam
Question #: 75
Topic #: 5
[All Professional Machine Learning Engineer Questions]

You have trained a DNN regressor with TensorFlow to predict housing prices using a set of predictive features. Your default precision is tf.float64, and you use a standard TensorFlow estimator;

estimator tf.estimator.DNNRegressor(

feature_columns[YOUR_LIST_OF_FEATURES],

hidden_units-[1024, 512, 256],

dropoutNone)

Your model performs well, but Just before deploying it to production, you discover that your current serving latency is 10ms @ 90 percentile and you currently serve on CPUs. Your production requirements expect a model latency of 8ms @ 90 percentile. You are willing to accept a small decrease in performance in order to reach the latency requirement Therefore your plan is to improve latency while evaluating how much the model's prediction decreases. What should you first try to quickly lower the serving latency?

Show Suggested Answer Hide Answer
Suggested Answer: D

Contribute your Thoughts:

Theodora
17 days ago
Quantization is the way to go here. Reducing the precision to tf.float16 should give us a nice latency boost without sacrificing too much model performance. And hey, at least we're not trying to run it on a Commodore 64, right?
upvoted 0 times
...
Zita
20 days ago
Haha, if I were the model, I'd be like 'Seriously? You want me to predict house prices in 8ms? What is this, the Flash's house?'
upvoted 0 times
Tenesha
2 days ago
A) Increase the dropout rate to 0.8 in_PREDICT mode by adjusting the TensorFlow Serving parameters
upvoted 0 times
...
...
Pamella
22 days ago
Switching to GPU serving is a good option, but it might be overkill for the 8ms latency requirement. I'd try the quantization approach first - reducing the precision to tf.float16 could give us a quicker win.
upvoted 0 times
Marilynn
3 days ago
C) Switch from CPU to GPU serving
upvoted 0 times
...
...
Catarina
2 months ago
Increasing the dropout rate to 0.8 in the _PREDICT mode seems like an interesting idea, but I'm not sure if that would actually improve the latency. Retraining the model with the higher dropout might work, but that could impact performance.
upvoted 0 times
Nu
10 days ago
A) Increase the dropout rate to 0.8 in _PREDICT mode by adjusting the TensorFlow Serving parameters
upvoted 0 times
...
...
Ngoc
2 months ago
But wouldn't applying quantization to the SavedModel also help reduce latency?
upvoted 0 times
...
Aliza
2 months ago
I agree with Loreen, that could help improve the latency.
upvoted 0 times
...
Loreen
2 months ago
I think we should try switching from CPU to GPU serving first.
upvoted 0 times
...

Save Cancel
a