Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Deploying ML Models in Google Cloud with Vertex AI

Deploying ML Models in Google Cloud with Vertex AI

Deploying ML Models in Google Cloud with Vertex AI

Olayinka Peter Oluwafemi

July 19, 2023
Tweet

More Decks by Olayinka Peter Oluwafemi

Other Decks in Technology

Transcript

  1. #MachineLearning Deploying ML Models in Google Cloud with Vertex AI

    Olayinka Peter Oluwafemi Snr ML, Youverify @olayinkapeter_ July 2023
  2. $whoami • ML engineering @ Youverify • ML GDE •

    PaLM API 󰙤 • ❤ TensorFlow, anime & peanut butter • Coordinates 󰗓 coffeewithpeter.com
  3. Data Neural Network diagram by Lia Koltyrina Data Why do

    we build ML models? Perfect Model (100% Accuracy)
  4. Data Neural Network diagram by Lia Koltyrina Data Why do

    we build ML models? Perfect Model (100% Accuracy) But what’s the use of a Perfect Model that’s not in Production?
  5. Some of the most famous tools used to deploy machine

    learning models Popular Model Deployment Tools Eiffel Tower by Travel-Fr
  6. TensorFlow Serve Torch Serve Each of them need some other

    tool to achieve end-to-end Pros and Cons of Our Favorites
  7. Enter Vertex AI Vertex AI combines data engineering, data science,

    and ML engineering workflows, enabling your teams to collaborate using a common toolset.
  8. Vertex AI is a one-stop shop machine learning platform that

    provides tools for every step of the machine learning workflow across different stages of machine learning development. It allows for Dataset Preparation (Gathering, Preprocessing and Version Control), Train (AutoML and Custom) and Deploy ML models and AI applications — all of these using the benefits of Google Cloud Introduction to Vertex AI Vertex AI: Introduction
  9. A-Z ML Workflow Simplified Simplified platform to manage end-to-end ML

    lifecycle, reduces complexity of managing separate components and services making development efficient. Pros #1
  10. Automated Model Deployment Makes deploying ML model at scale seamless.

    Packaging, versioning, A/B Testing, updates roll out with ease and without manual errors. Pros #2
  11. Integrated Experiments and Pipelines GC services like Cloud Blind, Kubeflow

    Pipelines etc. are integrated allowing for reproducibility, monitoring into unified pipelines and efficiency in the development process Pros #3
  12. Built-in Monitoring and Management Provides detailed metrics, logs and performance

    insights, allowing to track and evaluate performance, behavior, usage patterns in real-time Pros #4
  13. Scalability and Performance Handles large-scale workloads allowing to deploy and

    serve ML models handling heavy traffic and concurrent requests with Google Cloud Infrastructure Pros #5
  14. Collaboration Facilitates teamwork by providing shared project spaces, allowing multiple

    users/teams work on same project simultaneously and manage project collectively Pros #6
  15. Pre-trained Models and AutoML Integration AutoML capabilities and provides access

    to pre-trained models allowing to leverage existing models reducing extensive manual model development Pros #7
  16. Integrated Services Integrated with services like Cloud Storage, BigQuery, Dataflow

    to leverage processes like data ingestion, preprocessing, storage. Also advanced features like data encryption, access control, compliance certification, data governance etc. Pros #8
  17. Vertex AI Walkthrough Model Garden: Discover, Test, Customize and deploy

    Vertex AI and select open-source models (some pretrained) and assets Workbench: Jupyter Notebook based development environment, integrates Cloud storage and BigQuery to access and process data faster Pipelines: build and monitor pipelines to helps to automate, monitor and govern ML systems by orchestrating ML workflow in a serverless manner and store workflow’s artifacts using Vertex ML metadata Generative AI Studio: create, experiment with generative aI models. Test and customize Google’s LLMs Data: All data preparation takes place, can label, annotate and do a lot on the data Model Development: We can train ML models using either AutoML or Custom. After training, we can assess the model, optimize and even understand the signals behind the model’s predictions with “Explainable AI” Deploy and Use: Deploy Model to an endpoint to serve for online predictions using the API or the console. Includes all the physical resources and scalable hardware needed to scale the model for lower latency and online predictions. Undeployed model can also be used for Batch predictions using CLI, console UI or the SDK and the APIs. Each model can have multiple endpoints