Intelligent Product Essentials helps rapidly build products to deploy AI models at edge and leverage telemetry information along with learnings to further generate useful insights and optimize AI models in the cloud. It is essential to have an MLOps pipeline to streamline each stage of workflow, considering various elements, including ML model training workflows, geo-distributed deployment and source of data streams. This helps rapidly build machine learning models to support AI at the edge with appliances and products.
This document describes the various ML workflows, architectures, and MLOps pipelines for Intelligent Product Essentials using Vertex AI. The information in this document will help you do following:
As described in Hidden technical debt in ML systems, the ML code is only a small part of mature ML systems. In addition to the ML code and high-quality data, you need a way to put your ML processes into operation.
MLOps is a practice which helps companies to build, deploy and put your ML system into operation in a rapid, repeatable, and reliable manner. MLOps is an application of DevOps principles to ML systems. MLOps is an engineering culture and practice that's intended to unify ML system development (Dev) and ML system operation (Ops). The objective of MLOps is to provide a set of standardized processes and technology capabilities for building, deploying, and putting ML systems into operation rapidly and reliably.
There are many users within an organization who have a part to play in the MLOps life cycle, including a data scientist who dabbles in different aspects of building and validating models, to an ML engineer who is responsible for the model to work without issues to end users in production systems, to a software engineer who writes scalable distributed systems.
The previous diagram shows the following components:
The following overlaps occur:
The following sections discuss how MLOps can be implemented with Intelligent Products Essentials and Vertex AI.
The preceding diagram shows the following component and core MLOps user personas:
The subsequent sections explain the comprehensive spectrum of potential machine learning workflow architecture options for Intelligent Product Essentials in relation to diverse use cases.
The following diagram illustrates the various potential training and deployment scenarios for inferencing that are feasible in the context of any Intelligent Product Essential ML workflow.
These scenarios are based on the utilization of all the gathered information form Intelligent Products Essentials such as device data, device telemetry, and customer ownership data, that are stored within a data warehouse:
In the below section we will explore each of these points in more detail.
BigQuery ML is a model development service within BigQuery. With BigQuery ML, SQL users can train ML models directly in BigQuery without needing to move data or worry about the underlying training infrastructure. To learn more about the advantages of using BigQuery ML, see What is BigQuery ML?
To create a model in BigQuery, use the BigQuery ML CREATE MODEL statement. This statement is similar to the CREATE TABLE DDL statement. When you run a query that contains a CREATE MODEL statement, a query job is generated for you that processes the query.
For example for XGBoost model:
{CREATE MODEL | CREATE MODEL IF NOT EXISTS | CREATE OR REPLACE MODEL} model_name [INPUT(field_name field_type, …) OUTPUT(field_name field_type, …)] OPTIONS(MODEL_TYPE = 'XGBOOST', MODEL_PATH = string_value);
Create and train models with minimal technical knowledge and effort. To learn more about AutoML, see AutoML beginner's guide.
The above illustration elucidates the sequential stages integral to the construction of machine learning (ML) models. The process encompasses data preprocessing, feature engineering, model selection, hyperparameter tuning, evaluation, and deployment. Consequently, it transforms into a complex and time-consuming endeavor, even during the experimentation or evaluation phases of ML applications.
The AutoML product is designed to provide users with a graphical, codeless interface that guides them through the complete machine learning lifecycle, with significant automation and guardrails at each phase. This no code, low code interface facilitates the smooth definition of data schema and target, analysis of input features in the feature statistics dashboard, automated training of the model including automated feature engineering, model selection, and hyperparameter tuning. Furthermore, it enables the evaluation of model behavior prior to deployment in production and the deployment of the model with one click. Thus, it can help users reduce the time required for ML model creation from months to weeks or even days.
The following diagram demonstrates a Vertex AI custom ML pipeline. Create and train models at scale using any ML framework. To learn more about custom training on Vertex AI, see Custom training overview.
This is an example of the Vertex AI MLOps architecture with CI/CD showcasing the following components:
For each of the model training workflow, in addition to the deployment of Vertex AI online and batch prediction endpoint, based on the model type, the model can also be exported for edge or local deployment. For instance, BigQuery ML can be exported and deployed as an endpoint in both local and vertex endpoint environments, see exporting BQML Models for online predictions for more details.
Intelligent Product Essentials with MLOps will accelerate the product roadmaps and help you create engaging products for better customer experiences and feedback loops for various use cases. Refer to the Intelligent Product Essentials use case section for more details.
Authors:
Other contributors:
(Note: to see nonpublic LinkedIn profiles, sign in to LinkedIn.)