Section 4 · ~15% of Exam

Business Strategies for Generative AI

Covers the strategic and governance aspects of deploying generative AI in enterprise settings: implementation steps, securing AI systems with Google's SAIF framework, IAM and Security Command Center, and responsible AI principles including transparency, privacy, bias, fairness, accountability, and explainability.

Implementation Steps
SAIF Framework
Security (IAM, SCC)
Responsible AI Principles
Open in Colab Open Notebook in Colab
01

Gen AI Implementation Steps

Successfully implementing generative AI in an enterprise requires a structured approach. The exam tests your understanding of the recommended implementation framework — a phased methodology that minimizes risk while maximizing value.

Phase 1: Assess and Identify

Step Activities Key Deliverables
1. Define Objectives Identify business problems AI can solve, define success metrics Use case inventory, KPI definitions, ROI projections
2. Assess Readiness Evaluate data availability, team skills, infrastructure, governance Readiness assessment, gap analysis, training plan
3. Prioritize Use Cases Rank opportunities by impact, feasibility, and risk Prioritized roadmap, quick wins identified
4. Choose Technology Select models, platforms, and tools (Vertex AI, prebuilt products) Architecture decisions, vendor selection, cost estimates

Phase 2: Pilot, Iterate, and Scale

Step Activities Key Deliverables
5. Build Prototype Develop MVP using Vertex AI Studio, prompt engineering Working prototype, initial evaluation results
6. Pilot with Users Deploy to limited user group, collect feedback User feedback, quality metrics, safety analysis
7. Iterate & Optimize Refine prompts, add grounding/RAG, fine-tune if needed Improved model, optimized pipeline, cost analysis
8. Scale to Production Deploy at scale with monitoring, governance, and support Production deployment, SLAs, runbooks, monitoring dashboards
Exam Tip

The exam expects you to recommend the right implementation approach for the situation. For quick wins, recommend prebuilt products (Gemini for Workspace). For custom applications, recommend the prototype-pilot-scale approach using Vertex AI. Never recommend jumping straight to production without a pilot phase.

02

Securing AI Systems

AI systems introduce new security challenges beyond traditional software. Models can be manipulated through prompt injection, training data can be poisoned, and generated outputs can inadvertently leak sensitive information. Google's Secure AI Framework (SAIF) addresses these AI-specific security concerns.

Google's Secure AI Framework (SAIF)

SAIF is Google's comprehensive framework for securing AI systems throughout their lifecycle. It extends traditional security practices with AI-specific considerations and is a key topic on the exam.

SAIF Core Elements

🔒

1. Expand Security Foundations

Apply existing security best practices (least privilege, defense in depth, zero trust) to AI systems. Use VPC Service Controls, encryption, and audit logging for AI workloads.

🔍

2. Extend Detection & Response

Monitor AI systems for attacks: prompt injection, data exfiltration, model theft, and adversarial inputs. Extend SOC capabilities to cover AI-specific threats.

🔧

3. Automate Defenses

Use AI to enhance security itself. Automated input validation, output filtering, anomaly detection on model behavior, and continuous safety testing.

📖

4. Harmonize Platform Controls

Ensure consistent security controls across the AI platform. Standardize access policies, data handling, and compliance requirements for all AI workloads.

🚀

5. Adapt Controls for AI

Develop AI-specific security measures: input/output guardrails, model access controls, training data provenance tracking, and model versioning.

🌐

6. Contextualize AI Risks

Assess risks specific to each AI deployment context. A customer-facing chatbot has different risks than an internal code assistant. Tailor security controls accordingly.

Key Concept

SAIF is not just about technology — it is a risk management framework. The exam tests whether you understand that AI security requires both traditional security controls (IAM, encryption, networking) AND AI-specific measures (prompt injection defense, output filtering, model monitoring).

AI-Specific Threats

Threat Description Mitigation
Prompt Injection Malicious inputs that trick the model into ignoring instructions Input validation, system instructions, output filtering
Data Poisoning Corrupting training data to manipulate model behavior Data provenance, validation pipelines, access controls
Model Theft Extracting model weights or behavior through queries Rate limiting, API access controls, monitoring
Data Leakage Model revealing training data or PII in outputs DLP integration, output filtering, data anonymization
Adversarial Attacks Crafted inputs that cause misclassification or wrong outputs Adversarial training, robust model evaluation, safety filters
03

IAM for Generative AI

Identity and Access Management (IAM) is fundamental to securing AI workloads on Google Cloud. Vertex AI integrates with Cloud IAM to provide fine-grained access control over models, data, and AI resources.

Key IAM Roles for Vertex AI

Role Permissions Who Gets This
Vertex AI User Call model APIs, use Vertex AI Studio Developers, data scientists
Vertex AI Admin Full Vertex AI management (create, delete, configure) Platform administrators, ML engineers
Vertex AI Viewer Read-only access to models and resources Auditors, stakeholders
Service Account Automated access for applications and pipelines Production applications, CI/CD pipelines

Best practices for IAM in AI workloads:

  • Least privilege — grant only the minimum permissions needed for each role
  • Service accounts for apps — never use personal credentials for production workloads
  • VPC Service Controls — restrict API access to trusted networks
  • Audit logging — enable Cloud Audit Logs for all Vertex AI API calls
  • Separate environments — different projects for dev, staging, and production
04

Security Command Center

Security Command Center (SCC) is Google Cloud's centralized security and risk management platform. It provides visibility into security posture, vulnerability management, threat detection, and compliance monitoring for all Google Cloud resources, including AI workloads.

🔍

Asset Inventory

Discover and catalog all AI assets: Vertex AI models, endpoints, datasets, notebooks, and pipelines. Know what AI resources exist in your organization.

Vulnerability Detection

Identify misconfigurations in AI infrastructure: public endpoints, overpermissive IAM, unencrypted data stores, missing VPC controls.

🚨

Threat Detection

Detect suspicious activity on AI resources: unusual API access patterns, data exfiltration attempts, unauthorized model access.

📋

Compliance Monitoring

Verify AI workloads comply with organizational policies and regulatory requirements (HIPAA, SOC2, GDPR, FedRAMP).

05

Data Governance for Gen AI

Data governance is critical for AI because models are only as good as their data. Enterprise AI governance covers data lineage, quality, privacy, and access controls.

Governance Area Why It Matters for Gen AI Google Cloud Tools
Data Lineage Track where training/RAG data came from and how it was processed Dataplex, Data Catalog
Data Quality Ensure training and RAG data is accurate, complete, current Dataplex Data Quality, Dataprep
Data Privacy Prevent PII from appearing in training data or model outputs DLP API, data anonymization, CMEK
Data Access Control who can use which data for training and RAG IAM, VPC-SC, column-level security in BigQuery
Data Residency Keep data in required geographic regions for compliance Regional endpoints, data residency controls
Exam Focus

When the exam mentions a regulated industry (healthcare, finance, government), the answer almost always involves data governance, compliance controls, and data residency. Know that Vertex AI offers regional endpoints, CMEK encryption, VPC Service Controls, and DLP integration for these scenarios.

06

Responsible AI Principles

Google's Responsible AI principles are a cornerstone of the exam. Published in 2018, these principles guide how Google develops and deploys AI. The exam tests your understanding of each principle and how to apply them in practice.

Transparency

Transparency means being open about how AI systems work, what data they use, and what their limitations are. Users should know they are interacting with AI and understand the basis for AI-generated outputs.

  • Disclose when content is AI-generated
  • Provide model cards documenting model capabilities, limitations, and intended use
  • Share evaluation results and known failure modes
  • Enable source citations in RAG-based applications

Fairness and Bias

Fairness requires that AI systems do not create or reinforce unfair bias, particularly against protected groups. Bias in AI can arise from biased training data, biased model architecture choices, or biased evaluation criteria.

📊

Bias in Training Data

If training data underrepresents certain groups, the model performs worse for them. Mitigation: audit training data for representation, use balanced datasets.

📈

Bias in Outputs

Models may generate stereotypical or harmful content about certain groups. Mitigation: safety filters, red-teaming, diverse evaluation panels.

👥

Evaluation Bias

If evaluation only tests on majority cases, bias goes undetected. Mitigation: evaluate across demographics, use disaggregated metrics.

Privacy

Privacy in generative AI involves protecting personal information throughout the AI lifecycle — from training data collection to model outputs. Key considerations:

  • Training data privacy — ensure PII is removed or anonymized before training
  • Prompt privacy — user prompts may contain sensitive information; handle appropriately
  • Output privacy — prevent models from generating PII from memorized training data
  • Data processing agreements — ensure contractual protections for enterprise data
  • Right to be forgotten — mechanisms to remove individual data from model influence
Key Concept

Google Cloud's data processing commitment for Vertex AI: customer data submitted through Vertex AI APIs is not used to train Google's foundation models. This is a critical selling point for enterprise adoption and an important exam topic.

Accountability

Accountability means clear ownership and responsibility for AI system behavior. Organizations must establish:

  • Clear ownership — designated teams responsible for each AI system
  • Audit trails — logs of all model decisions for post-hoc review
  • Escalation paths — procedures for addressing AI failures or harmful outputs
  • Regular reviews — periodic assessment of AI system performance and impact
  • Incident response — plans for handling AI-related security or safety incidents

Explainability

Explainability enables understanding of why an AI system produced a particular output. For generative AI, explainability includes:

  • Source attribution — RAG-based systems can cite the documents they used
  • Grounding metadata — Vertex AI provides grounding scores and source links
  • Confidence scores — indicating the model's certainty in its response
  • Reasoning traces — chain-of-thought prompting makes reasoning visible
07

Google's Responsible AI in Practice

Google applies responsible AI through concrete technical mechanisms and organizational processes:

🛡

Safety Filters

Built-in content safety filters on all Gemini models. Block harmful content across categories: hate speech, harassment, dangerous content, sexually explicit material. Configurable thresholds.

📍

Model Cards

Documentation for each model describing its capabilities, limitations, intended use, training data, evaluation results, and ethical considerations.

🔮

Red Teaming

Adversarial testing where security experts try to make models produce harmful outputs. Results feed back into model safety improvements.

📋

AI Principles Review

Internal review process where new AI applications are evaluated against Google's AI Principles before launch. Includes ethics review for high-risk applications.

🌟

SynthID

Digital watermarking technology that embeds imperceptible marks in AI-generated images and text. Enables detection of AI-generated content for transparency.

📈

Vertex AI Evaluation

Built-in evaluation tools for measuring model safety, fairness, and quality. Compare models on safety benchmarks before deploying to production.

Exam Tip

Google's seven AI principles state that AI should: (1) be socially beneficial, (2) avoid creating or reinforcing unfair bias, (3) be built and tested for safety, (4) be accountable to people, (5) incorporate privacy design principles, (6) uphold high standards of scientific excellence, (7) be made available for uses that accord with these principles. The exam may reference these directly.

08

Measuring Gen AI ROI

Leaders need to demonstrate the business value of generative AI investments. The exam tests your understanding of how to measure and communicate AI ROI.

Metric Category Specific Metrics Example
Productivity Time saved, tasks automated, throughput increase 50% reduction in email drafting time with Gemini for Workspace
Cost Reduction Labor costs saved, operational efficiency gains 30% reduction in Tier 1 support costs with AI chatbot
Revenue Impact New revenue streams, conversion rates, customer lifetime value 15% increase in conversion with personalized AI recommendations
Quality Error rates, customer satisfaction, response accuracy 20% improvement in customer satisfaction scores
Innovation New capabilities, time-to-market, competitive advantage Launched AI-powered product feature 3 months ahead of competitors
Key Concept

When calculating ROI, include both direct costs (API usage, compute, development) and indirect costs (training, change management, governance). Also account for risk costs (potential reputational damage from AI failures, compliance penalties). The best AI projects start with high-value, low-risk use cases.

Change Management

Successful AI adoption requires organizational change management:

  • Executive sponsorship — C-level support for AI initiatives
  • Training and upskilling — teach employees to work with AI tools effectively
  • Clear communication — explain how AI augments (not replaces) human work
  • Governance framework — policies for responsible AI use
  • Center of excellence — dedicated team to guide AI adoption across the organization
  • Feedback loops — mechanisms for users to report issues and suggest improvements