☁️ Google Cloud Professional ML Engineer

Machine Learning
Engineer Path

Master ML engineering on Google Cloud — from BigQuery ML and Vertex AI to production MLOps pipelines. 20 deep-dive study guides mapped to all 6 exam sections.

Begin with Course 01 →
20
Study Guides
6
Exam Sections
20
Notebooks
🌐
Open Source
All Study Guides
Each guide covers key concepts with hands-on examples.
Section 1 · Architecting Low-Code ML Solutions
COURSE 01 · SECTION 1
Introduction to AI and ML on Google Cloud
AI foundations, Vertex AI overview, and generative AI capabilities on Google Cloud Platform.
Vertex AIAI FoundationsGenAI on GCPML Services
COURSE 02 · SECTION 1
Prepare Data for ML APIs
Using Vision, NLP, Speech, and Translation APIs alongside Dataflow and Dataprep for data preparation.
Vision APINLP APIDataflowDataprep
COURSE 03 · SECTION 1
Working with Notebooks in Vertex AI
Colab Enterprise, Vertex AI Workbench, and JupyterLab environments for ML development on GCP.
Colab EnterpriseWorkbenchJupyterLabManaged Notebooks
COURSE 04 · SECTION 1
Create ML Models with BigQuery ML
Classification, regression, forecasting, and matrix factorization models built directly in SQL with BigQuery ML.
BigQuery MLClassificationForecastingMatrix Factorization
COURSE 05 · SECTION 1
Engineer Data for Predictive Modeling with BigQuery ML
ETL pipelines, Dataprep, Dataflow integration, and feature engineering techniques for BigQuery ML models.
ETLDataprepDataflowFeature Engineering
Section 2 · Collaborating to Manage Data and Models
COURSE 06 · SECTION 2
Feature Engineering
Feature Store, transforming raw data into features, TensorFlow Transform, and feature crosses for ML models.
Feature StoreTF TransformFeature CrossesRaw Data
Section 3 · Scaling Prototypes into ML Models
COURSE 07 · SECTION 3
TensorFlow on Google Cloud
The TF ecosystem, input pipelines with tf.data, Keras model building, and distributed training strategies on GCP.
TensorFlowtf.dataKerasDistributed Training
COURSE 08 · SECTION 3
Production ML Systems
ML system architecture, adaptable and high-performance systems, hybrid deployment patterns, and troubleshooting production ML.
ArchitectureHigh-PerformanceHybrid MLTroubleshooting
Section 4 · Serving and Scaling Models
COURSE 09 · SECTION 4
MLOps Getting Started
ML operations fundamentals and getting started with Vertex AI MLOps for model lifecycle management.
MLOpsVertex AIModel LifecycleCI/CD for ML
COURSE 10 · SECTION 4
MLOps with Vertex AI: Manage Features
Deep dive into Vertex AI Feature Store for feature management, serving, and sharing across ML workflows.
Feature StoreFeature ServingFeature SharingVertex AI
Section 5 · Automating and Orchestrating ML Pipelines
COURSE 11 · SECTION 5
Introduction to Generative AI
Generative AI fundamentals -- what it is, how it works, and its applications in the ML pipeline ecosystem.
GenAI BasicsFoundation ModelsUse CasesGCP GenAI
COURSE 12 · SECTION 5
Introduction to Large Language Models
Core LLM concepts -- architecture, training paradigms, prompting techniques, and fine-tuning for enterprise use.
LLM ArchitecturePromptingFine-TuningPaLM / Gemini
COURSE 13 · SECTION 5
MLOps for Generative AI
GenAI-specific MLOps patterns -- model versioning, prompt management, evaluation pipelines, and deployment strategies for generative models.
GenAI MLOpsPrompt MgmtModel VersioningEvaluation
COURSE 14 · SECTION 5
MLOps with Vertex AI: Model Evaluation
Model evaluation techniques for both traditional ML and generative AI models using Vertex AI evaluation tools.
Model EvaluationVertex AIMetricsGenAI Eval
COURSE 15 · SECTION 5
ML Pipelines on Google Cloud
Building and orchestrating ML pipelines with TFX, Kubeflow Pipelines, Cloud Composer, and MLflow on GCP.
TFXKubeflowCloud ComposerMLflow
Section 6 · Monitoring ML Solutions
COURSE 16 · SECTION 6
Build and Deploy ML Solutions on Vertex AI
AutoML, custom training pipelines, online and batch predictions, and end-to-end deployment on Vertex AI.
AutoMLCustom TrainingPredictionsVertex Pipelines
COURSE 17 · SECTION 6
Build Generative AI Applications
Prompting strategies, retrieval-augmented generation, and building production GenAI applications on Google Cloud.
PromptingRAGGenAI AppsVertex AI Search
COURSE 18 · SECTION 6
Responsible AI: Fairness and Bias
Fairness metrics, bias detection and mitigation techniques, and building equitable ML systems on Google Cloud.
Fairness MetricsBias DetectionMitigationWhat-If Tool
COURSE 19 · SECTION 6
Responsible AI: Interpretability and Transparency
Model explainability techniques including SHAP values, feature attributions, and the What-If Tool for transparent ML.
ExplainabilitySHAPWhat-If ToolFeature Attribution
COURSE 20 · SECTION 6
Responsible AI: Privacy and Safety
PII handling, differential privacy, federated learning, and safety frameworks for production ML systems.
PIIDifferential PrivacyFederated LearningSafety
Glossary
BigQuery ML
Build and deploy ML models directly in BigQuery using SQL. Supports classification, regression, forecasting, clustering, matrix factorization, and imported TensorFlow models.
Vertex AI
Google Cloud's unified ML platform. Provides tools for dataset management, training (AutoML and custom), model deployment, pipelines, feature store, experiments, and model monitoring.
Feature Store
A centralized repository for organizing, storing, and serving ML features. Ensures consistency between training and serving, supports point-in-time lookups, and enables feature sharing across teams.
AutoML
Automated machine learning that handles model architecture search, hyperparameter tuning, and feature engineering. Available in Vertex AI for tabular, image, text, and video data.
TFX (TensorFlow Extended)
An end-to-end platform for deploying production ML pipelines. Components include ExampleGen, StatisticsGen, SchemaGen, Transform, Trainer, Tuner, Evaluator, and Pusher.
Kubeflow Pipelines
A platform for building and deploying portable, scalable ML workflows on Kubernetes. Supports DAG-based pipeline definitions with reusable components and experiment tracking.
TPU
Tensor Processing Unit. Google's custom ASIC designed for high-throughput ML training and inference. Available in Cloud TPU pods for distributed training at massive scale.
MLOps
Machine Learning Operations. The practice of applying DevOps principles to ML systems -- continuous integration, delivery, and training (CI/CD/CT) for ML models in production.
Model Registry
A central catalog for managing ML model versions, metadata, lineage, and deployment status. Vertex AI Model Registry tracks models from training through deployment.
Dataflow
A fully managed stream and batch data processing service based on Apache Beam. Used for ETL pipelines, data transformation, and real-time feature computation for ML.
Cloud Composer
A managed Apache Airflow service for orchestrating complex workflows. Commonly used to schedule and coordinate ML pipeline steps, data ingestion, and model retraining.
TF Transform
A TFX library for preprocessing data using TensorFlow. Ensures the same transformations are applied during both training and serving, preventing training-serving skew.
Vertex AI Pipelines
A serverless ML workflow orchestration service. Supports both TFX and Kubeflow Pipelines SDK, with built-in lineage tracking and integration with other Vertex AI services.
Model Monitoring
Continuous tracking of model performance in production. Detects data drift, concept drift, and feature skew by comparing serving data distributions against training baselines.
Data Drift
When the statistical properties of production input data diverge from training data over time. A primary cause of model degradation that triggers retraining in MLOps pipelines.
Training-Serving Skew
Differences between how data is processed during training versus serving. Causes prediction errors in production. TF Transform and Feature Store are key mitigations.
Hyperparameter Tuning
The process of finding optimal model configuration (learning rate, layers, batch size). Vertex AI Vizier provides automated Bayesian optimization for hyperparameter search.
Vertex AI Experiments
A managed experiment tracking service for comparing model training runs. Logs metrics, parameters, and artifacts to help identify the best-performing model configurations.
Explainable AI
Techniques for understanding model predictions. Vertex AI provides feature attributions using Sampled Shapley, Integrated Gradients, and XRAI for image models.
SHAP
SHapley Additive exPlanations. A game-theoretic approach to model interpretability that assigns each feature an importance value for a particular prediction.
What-If Tool
An interactive visual tool for exploring ML model behavior. Allows probing model performance, testing fairness across subgroups, and investigating individual predictions without code.
Differential Privacy
A mathematical framework for quantifying privacy guarantees in data analysis and ML. Adds calibrated noise to ensure individual records cannot be identified from model outputs.
Federated Learning
A distributed ML approach where models are trained across multiple devices without centralizing data. Each device trains locally and only shares model updates, preserving data privacy.
Vertex AI Endpoints
Managed serving infrastructure for deployed models. Supports online (real-time) and batch predictions, traffic splitting for A/B testing, autoscaling, and model versioning.
MLflow
An open-source platform for the ML lifecycle. Provides experiment tracking, model packaging, registry, and deployment. Can be integrated with Vertex AI for hybrid workflows.
Colab Enterprise
A managed Jupyter notebook environment in Google Cloud with enterprise security, VPC-SC support, and integration with BigQuery, GCS, and Vertex AI services.
Dataprep
An intelligent data preparation service (by Trifacta) for visually exploring, cleaning, and transforming structured and unstructured data for ML. Runs on Dataflow under the hood.
Feature Cross
A synthetic feature created by combining two or more features. Enables linear models to learn non-linear relationships. Common in BigQuery ML and wide-and-deep models.
Distributed Training
Training ML models across multiple machines or accelerators. Strategies include data parallelism (split data) and model parallelism (split model). Supported natively in Vertex AI.
RAG
Retrieval-Augmented Generation. Combines information retrieval with generative models to ground responses in external knowledge. Vertex AI Search and Conversation enables managed RAG.