Google Professional-Machine-Learning Exam

Questions Number: 32 out of 339 Questions
9.44%

Question 32
You developed an ML model with AI Platform, and you want to move it to production. You serve a few thousand queries per second and are experiencing latency issues. Incoming requests are served by a load balancer that distributes them across multiple Kubeflow CPU-only pods running on Google Kubernetes Engine
(GKE). Your goal is to improve the serving latency without changing the underlying infrastructure. What should you do?







Previous Questions Next Questions



Premium Version