Skip to content

Google Cloud Platform Vertex AI end-to-end workflows for machine learning operations

License

Notifications You must be signed in to change notification settings

statmike/vertex-ai-mlops

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

tracker

GitHub logo
View on
GitHub





Vertex AI for Machine Learning Operations

2024 UPDATE: This repository is evolving from end-to-end workflows for various frameworks into an MLOps focused approach for development of predictive and generative AI operations. The new approach is being developed in the MLOps folder. Once it nears completion, the content in this repository will be rearranged into the following structure:

  • MLOps
    • Pipelines
    • Experiments
    • Feature Store
    • Model Monitoring
    • ...
  • Applied Examples
    • Forecasting
    • GenAI
    • ...
  • Framework Workflows
    • BigQuery ML
    • TensorFlow
    • scikit-learn
    • ...
  • ...

This is the original readme from prior to the shift in this repository. After the content rearrangement is complete and this information is incorporate above it will be removed.

👋 I'm Mike

I want to share and enable Vertex AI from Google Cloud with you. The goal here is to share a comprehensive set of end-to-end workflows for machine learning that each cover the range of data to model to serving and managing - even automating the flow. Regardless of your data type, skill level or framework preferences you will find something helpful here. You can even ask for what you need and I might be able to work it into updates!

Click to watch on YouTube

Click here to see current playlist for this repository


Tracking

To better understand which content is most helpful to users, this repository uses tracking pixels in each markdown (.md) and notebook (.ipynb) file. No user or location data is collected. The only information captured is that the content was rendered/viewed which gives us a daily count of usage. Please share any concerns you have with this in repositories discussion board and I am happy to also provide a branch without the tracking.

A script is provided to remove this tracking from your local copy of this repository in the file pixel_remove.py in the folder pixel. This readme also has the complete code for creating the tracking in case you want to use replicate it or just understand it in greater detail.


Approach Used In This Repository

This repository is presented as workflows using, primarily, interactive python notebooks .ipynb. Why? These are easy to review, share, and move. They contain elements for both code and narrative. The narrative can be written with plain text, Markdown and/or HTML which makes providing visual explanations easy. This reinforces the goal of this repository: information that is easily accessible, portable, and great for starting points in your own work.

In notebooks, execution is driven from the locally attached compute. In this repository that means the Python code is currently running in the notebooks compute. The code in this repository heavily leans on orchestrating services in GCP rather than doing data compute in the local environment to the notebook. That means these notebooks are designed to run on minimal machine sizes, like n1-standard2 even. The heavy work of training and serving is done on Vertex AI, BigQuery, and other Google Cloud services. You will even find notebooks that author code, and then deploy the code in services like Vertex AI Custon Training and Vertex AI Pipelines.

There are sections that use other languages, like R, as well as creating files that are external to the notebooks: dockerfile, .py scripts and modules, etc.

The code in this repository is opinionated. It is not completely production ready as well as not simply ad-hoc exploration. It aims to the right of the continum of exploration to deployment: 'hello-world' to CI/CD/CT. In our data science daily work we might think of the process as:

In explore, everything is code as you go. At some point in this exploration ideas find value and need to be developed.

In develop, the approach is usually something like:

  • make it work
    • get a working end to end flow
  • clean it up
    • revisit the code and remove parts that are no longer needed and reorder based on what is learned
  • generalize it
    • parameterize
    • use functions
    • control flow: start using logic to check for out of bound conditions
  • optimize it
    • better use of data structures to handle data usage during execution
    • consider execution timing and optimize for the simoultaneous goal of readability (= maintainability) and compute time

In many cases, getting from development to deployment is simple:

  • schedule a notebook - a lot like skipping the develop stage
  • deploy a pipeline
  • create a cloud function

But, inevitably, as a workflow proves value it requires more effort before you deploy:

  • error handling
  • unit testing
  • move from specialized code to generalized code:
    • use classes
    • control environment handling

So where does the code in the repository fall? In the late develop phase with strong readability and adaptibility.


Table of Contents


Considerations

Data Type

  • Tables: Tabular, structured data in rows and columns
  • Language: Text for translation and/or understanding
  • Vision: Images
  • Video

Convenience Level

  • Use Pre-Trained APIs
  • Automate building Custom Models
  • End-to-end Custom ML with core tools in the framework of your choice

Framework Preferences


Overview

This is a series of workflow demonstrations that use the same data source to build and deploy the same machine learning model with different frameworks and automation. These are meant to help get started in understanding and learning Vertex AI and provide starting points for new projects.

The demonstrations focus on workflows and don't delve into the specifics of ML frameworks other than how to integrate and automate with Vertex AI. Let me know if you have ideas for more workflows or details to include!

To understand the contents of this repository, the following charts uncover the groupings of the content.

Direction

Pre-Trained APIs

Pre-Trained Models AutoML
Data Type Pre-Trained Model Prediction Types Related Solutions

Text

Cloud Translation API
Detect, Translate
Cloud Text-to-Speech

AutoML Translation

Cloud Natural Language API
Entities (Identify and label), Sentiment, Entity Sentiment, Syntax, Content Classification
Healthceare Natural Language API

AutoML Text

Image

Cloud Vision API
Crop Hint, OCR, Face Detect, Image Properties, Label Detect, Landmark Detect, Logo Detect, Object Localization, Safe Search, Web Detect

Document AI

Visual Inspection AI

AutoML Image

Audio

Cloud Media Translation API
Real-time speech translation
Cloud Speech-to-Text

Video

Cloud Video Intelligence API
Label Detect*, Shot Detect*, Explicit Content Detect*, Speech Transcription, Object Tracking*, Text Detect, Logo Detect, Face Detect, Person Detect, Celebrity Recognition
Vertex AI Vision

AutoML Video

AutoML

AutoML
Data Type
AutoML
Prediction Types

Table

AutoML Tables
Classification
Binary
Multi-class
Regression
Forecasting

Image

AutoML Image
Classification
Single-label
Multi-label
Object Detection

Video

AutoML Video
Classification
Object Detection
Action Recognition

Text

AutoML Text
Classification
Single-label
Multi-label
Entity Extraction
Sentiment Analysis

Text

AutoML Translation
Translation

With Training Data

This work focuses on cases where you have training data:

Overview
AutoML BigQuery ML Vertex AI Forecasting with AutoML, BigQuery ML, OSS Prophet

Vertex AI For ML Training

       


Vetex AI

Vetex AI is a platform for end-to-end model development. It consist of core components that make the processes of MLOps possible for design patterns of all types.

Components         Console


Interacting with Vertex AI

Many Vertex AI resources can be viewed and monitored directly in the GCP Console. Vertex AI resources are primarily created, and modified with the Vertex AI API.

The API is accessible from:

The notebooks in this repository primarily use the Python client aiplatform. There is occasional use of aiplatform.gapic, aiplatform_v1 and aiplatform_v1beta1.

For the full details on the APIs versions and layers and how/when to use each, see this helpful note.

Install the Vertex AI Python Client

pip install google-cloud-aiplatform

Example Usage: Listing all Models in Vertex AI Model Registry

PROJECT = 'statmike-mlops-349915'
REGION = 'us-central1'

# List all models for project in region with: aiplatform
from google.cloud import aiplatform
aiplatform.init(project = PROJECT, location = REGION)

model_list = aiplatform.Model.list()

Setup

The demonstrations are presented in a series of notebooks that are best run in JupyterLab. These can be reviewed directly in this repository on GitHub or cloned to your Jupyter instance on Vertex AI Workbench Instances.

Option 1: Review And Use Individual Files

Select the files and review them directly in the browser or IDE of your choice. This can be helpful for general understanding and selecting sections to copy/paste to your project. Some options to get a local copy of this repositories content:

  • use git: git clone https://github.com/statmike/vertex-ai-mlops
  • use wget to copy individual files directly from GitHub:
    • Go to the notebook on GitHub.com and right-click the download link. Then select copy link address.
    • Alternatively, click the Raw button on GitHub and then copy the URL that loads.
    • Run the following from a notebook cell or directly from a terminal (without the !). Note the slightly different address that points directly to raw content on GitHub.
      • !wget "https://raw.githubusercontent.com/statmike/vertex-ai-mlops/main/<path and filename>.ipynb"
  • Use Colab (and soon Vetex AI Enterprise Colab) to open the notebooks. Many of the notebooks have section at the top with buttons for opening directly in Colab. Some notebooks don't yet have this feature and some use local Docker which is not available on Colab.

Option 2: Run These Notebooks in a Vertex AI Workbench based Notebook

TL;DR

In Google Cloud Console, Select/Create a Project then go to Vertex AI > Workbench > Instances

  • Create a new notebook and Open JupyterLab
  • Clone this repository using Git Menu, Open and run 00 - Environment Setup.ipynb
  1. Create a Project
    1. Link, Alternatively, go to: Console > IAM & Admin > Manage Resources
    2. Click "+ Create Project"
    3. Provide: name, billing account, organization, location
    4. Click "Create"
  2. Enable the APIs: Vertex AI API and Notebooks API
    1. Link
      1. Alternatively, go to:
        1. Console > Vertex AI, then enable API
        2. Then Console > Vertex AI > Workbench, then enable API
  3. Create A Notebook with Vertex AI Workbench Instances:
    1. Go to: Console > Vertex AI > Workbench > Instances - direct link
    2. Create a new instance - instructions
    3. Once it is started, click the Open JupyterLab link.
    4. Clone this repository to the JupyterLab instance:
      1. Either:
        1. Go to the Git menu and choose Clone a Repository
        2. Choose the Git icon on the left toolbar and click Clone a Repository
      2. Provide the Clone URI of this repository: https://github.com/statmike/vertex-ai-mlops.git
      3. In the File Browser you will now have the folder "vertex-ai-mlops" that contains the files from this repository
  4. Setup the Notebook Environment for these workflows
    1. Open the notebook vertex-ai-mlops/00 - Environment Setup
    2. Follow the instructions and run the cells

Resources on these items:


Helpful Sections

  • Learning Machine Learning
    • I often get asked "How do I learn about ML?". There are lots of good answers. ....
  • Explorations
    • This is a series of projects for exploring new, new-to-me, and emerging tools in the ML world!
  • Tips
    • Tips for using the repository and notebooks with examples of core skills like building containers, parameterizing jobs and interacting with other GCP services. These tips help with scaling jobs and developing them with a focus on CI/CD.

More Resources Like This Repository

This is my personal repository of demonstrations I use for learning and sharing Vertex AI. There are many more resources available. Within each notebook I have included a resources section and a related training section.

About

Google Cloud Platform Vertex AI end-to-end workflows for machine learning operations

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •