Upgrade to Pro — share decks privately, control downloads, hide ads and more …

DevOps and AI

DevOps and AI

Henk Boelman

March 29, 2021
Tweet

More Decks by Henk Boelman

Other Decks in Technology

Transcript

  1. Sophisticated pretrained models To simplify solution development Azure Databricks Machine

    Learning VMs Popular frameworks To build advanced deep learning solutions TensorFlow Keras Pytorch Onnx Azure Machine Learning Language Speech … Azure Search Vision On-premises Cloud Edge Productive services To empower data science and development teams Powerful infrastructure To accelerate deep learning Flexible deployment To deploy and manage models on intelligent cloud and edge Machine Learning on Azure Cognitive Services
  2. DevOps is the union of people, process, and products to

    enable continuous delivery of value to your end users. “ ”
  3. Ask a sharp question Collect the data Prepare the data

    Select the algorithm Train the model Use the answer The data science process
  4. Azure Machine Learning A fully-managed cloud service that enables you

    to easily build, deploy, and share predictive analytics solutions.
  5. What is Azure Machine Learning? Set of Azure Cloud Services

    Python SDK Prepare Data Build Models Train Models Manage Models Track Experiments Deploy Models That enables you to:
  6. Datasets – registered, known data sets Experiments – Training runs

    Pipelines – Training workflows Models – Registered, versioned models Endpoints: Real-time Endpoints – Deployed model endpoints Pipeline Endpoints – Training workflow endpoints Compute – Managed compute Environments – defined training and inference environments Datastores – Connections to data Azure Machine Learning
  7. Azure Machine Learning Pipelines Workflows of steps that can use

    Data Sources, Datasets and Compute targets Unattended runs Reusability Tracking and versioning
  8. Azure Pipelines Orchestration for Continuous Integration and Continuous Delivery Gates,

    tasks and processes for quality Integration with other services Trigger on code and non-code events
  9. Create a pipeline step Input Output Runs a script on

    a Compute Target in a Docker container. Parameters
  10. Create a pipeline Dataset of Simpsons Images Prepare data Train

    the Model with PyTorch Processed dataset model Register the model Blob Storage Account Model Management
  11. Jupyter Notebook Compute Target Docker Image Data store 1. Snapshot

    folder and send to experiment 2. create docker image 3. Deploy docker and snapshot to compute 4. Mount datastore to compute 6. Stream stdout, logs, metrics 5. Launch the script 7. Copy over outputs Experiment
  12. Code and comments only (not Jupyter output) Plus every part

    of the pipeline And Infrastructure and dependencies And maybe a subset of data Source Control
  13. Everything should be in source control! Except your training data

    which should be a known, shared data source
  14. Triggered on code change Refresh and execute AML Pipeline Code

    quality, linting, and unit testing Pull request process Continuous Integration
  15. Trigger on model registration Deploy to test and staging environments

    Run integration and load tests Control: rollout, feature flags, A/B testing Continuous Delivery