Overview of Vertex AI

Vertex AI is a unified, open platform for building, deploying, and scaling generative AI and machine learning (ML) models and AI applications. It provides access to the Model Garden, featuring a curated catalog of over 200 models—including Google's foundation models (such as Gemini) and a comprehensive selection of partner and open models—along with the underlying TPU/GPU infrastructure. Vertex AI supports cutting-edge GenAI workflows as well as AI inference workflows for MLOps. It offers end-to-end MLOps tools and enterprise-grade controls for governance, security, and compliance.

Key capabilities of Vertex AI

Vertex AI includes tools and services that support generative AI as well as AI inference and machine learning workflows.

Generative AI capabilities

Vertex AI brings together a comprehensive toolset with Google's advanced foundation models tools that you can use to build production-ready generative AI agents and applications, as follows:

  • Prompting: Start with prompt design in Vertex AI Studio. Vertex AI Studio includes tools for prompt design and model management that you can use to prototype, build, and deploy generative AI applications.

  • Models: Vertex AI Model Garden is a centralized hub containing over 200 enterprise-ready models from Google, leading third-party partners (such as Anthropic's Claude), and popular open-source options (such as Llama).

    This selection of models includes the following:

    • Google's foundational generative AI models:

      • Gemini: Multimodal capabilities for text, images, video, and audio; and thinking capabilities for models, such as Gemini 3 Flash and Gemini 3 Pro (with Nano Banana).
      • Imagen on Vertex AI: Generate and edit images.
      • Veo on Vertex AI: Generate videos from text and images.
    • Partner and open source models: Access a curated selection of leading models such as Anthropic's Claude, Mistral AI models, and Llama with superior price-performance. These models are available as fully managed model as a service (MaaS) APIs.

  • Model customization: Tailor models to your business to create unique AI assets. This ranges from Grounding with your enterprise data or Google Search to reduce hallucinations, to using Vertex AI Training for Supervised Fine-Tuning (SFT) or Parameter-Efficient Fine-Tuning (PEFT) of models like Gemini. For more information about model customization, see Introduction to tuning.

  • Generative AI Evaluations: Objectively assess and compare model and agent performance with Vertex AI GenAI Evaluation capabilities. Ensure safety and compliance by deploying runtime defense features like Model Armor to proactively inspect and protect against emergent threats, such as prompt injection and data exfiltration.

  • Agent builders: Vertex AI Agent Builder is a full-stack agentic transformation system that helps you create, manage, and deploy AI agents. Use the open-source Agent Development Kit (ADK) to build and orchestrate agents, and then deploy them to the managed, serverless Vertex AI Agent Engine for use at scale in production. Each agent is assigned an Agent Identity (Identity and Access Management Principal) for security and a clear audit trail.

  • Access External Information: Enhance model responses by connecting to reliable sources with Grounding, interacting with external APIs using Function Calling, and retrieving information from knowledge bases with RAG.

  • Responsible AI and Safety: Use built-in safety features to block harmful content and ensure responsible AI usage.

For more information about Generative AI on Vertex AI, see the Generative AI on Vertex AI documentation.

AI inference capabilities

Vertex AI provides tools and services that map to each stage of the ML workflow:

  1. Data preparation: Collect, clean, and transform your data.

  2. Model training: Train your ML model.

    • Choose between AutoML for code-free training or Custom training for full control.
    • Manage and compare training runs using Vertex AI Experiments.
    • Register trained models in the Vertex AI Model Registry.
    • Vertex AI Training offers both serverless training and training clusters.
      • Use Vertex AI serverless training to run your custom training code on-demand in a fully managed environment. See the [Vertex AI serverless training overview][serverless].
      • Use Vertex AI training clusters for large jobs that need assured capacity on dedicated, reserved accelerator clusters. See Vertex AI training clusters overview.
    • Use Ray on Vertex AI to scale Python and ML workloads with the open-source Ray framework on a managed, interactive cluster. See Ray on Vertex AI overview.
    • Use Vertex AI Vizier to adjust model hyperparameters in complex ML models.
  3. Model evaluation and iteration: Assess and improve model performance.

  4. Model serving: Deploy and get inferences from your model.

  5. Model monitoring: Track deployed model performance over time.

Illustration of the machine learning workflow

MLOps Tools

Automate, manage, and monitor your ML projects:

What's next