AI Factory in Hybrid Manager v1.3

The AI Factory workload in Hybrid Manager brings scalable AI, machine learning, and Gen AI capabilities to your Hybrid Manager platform. It enables you to operationalize AI across your Hybrid Manager–managed clusters and data — with deep integration across Postgres, vector search, model serving, and AI assistants.

With AI Factory in Hybrid Manager, you can:

  • Deploy Gen AI assistants and agents for internal or external-facing use
  • Serve AI models at scale with integrated KServe-powered inferencing and GPU acceleration
  • Create Knowledge Bases and perform Retrieval-Augmented Generation (RAG) with vector search
  • Manage your model library and deploy trusted models under Hybrid Manager governance
  • Integrate AI features directly into your applications and data pipelines

Hub quick links: AI Factory HubGen AI (hub)Model Serving

For in-depth architecture and core concepts, see the AI Factory Hub.


Example AI solutions you can build

AI Factory in Hybrid Manager enables solutions across many domains:

  • Enterprise search and knowledge assistants Build RAG-based assistants that integrate with corporate documents, databases, and intranet content.

  • Customer support chatbots Deploy assistants powered by your own data and domain-specific models, combining conversational AI with Knowledge Bases.

  • AI-driven data apps Expose AI-powered endpoints for apps — semantic search, recommendations, similarity search, or natural language querying.

  • Operational AI for internal tools Build AI agents and tools to assist with DevOps, customer success, HR automation, sales enablement, and more.

  • Domain-specific model serving Serve proprietary or fine-tuned models (LLMs, embedding models, ranking models) as scalable inference services within your Hybrid Manager cluster.

You can start small — with a single assistant or model endpoint — and scale to full AI-powered applications using the AI Factory architecture.


Learning paths

Follow structured learning paths to master AI Factory in Hybrid Manager:


Use cases and solutions

We provide detailed guidance and patterns to help you build solutions with AI Factory:


AI Factory in Hybrid Manager workloads

Hybrid Manager supports a full set of AI Factory capabilities, tightly integrated into its control plane:

Gen AI workloads

Model management and serving

Vector Engine

  • Vector Engine — Integrated vector search and similarity capabilities with Postgres

Hybrid Manager learn content

In addition to AI Factory content in the hub, Hybrid Manager provides additional Learn content for Hybrid Manager–specific usage:


How AI Factory fits within Hybrid Manager

Hybrid Manager is the control plane for your databases, analytics, and AI workloads. AI Factory complements other HM capabilities:

  • Databases: Provision and manage Postgres clusters; AI Factory connects to these for vector search, RAG, and application endpoints.
  • Analytics: Offload data to Iceberg/Delta and query via SQL; AI Factory’s Pipelines and Vector Engine integrate with these datasets for retrieval.
  • Operations: Use HM observability and governance to monitor Gen AI, Pipelines, and model serving alongside databases and analytics.

Prerequisites for AI Factory in HM:

  • A Hybrid Manager project with AI Factory enabled and required entitlements for Gen AI, Pipelines, Vector Engine, and Model Serving.
  • Access to GPU resources where needed for serving. See GPU resource management.

Get started

To begin building with AI Factory in Hybrid Manager:

AI Factory in Hybrid Manager gives you a scalable, secure platform to operationalize AI across your hybrid data estate.


Getting started

Getting started

Quick orientation to AI Factory on Hybrid Manager with pointers to hub concepts and how-tos.

Architecture

Technical architecture of AI Factory components within Hybrid Manager's Kubernetes infrastructure

Sovereign AI

Applying Sovereign AI principles with AI Factory in HM.

Use Cases & Personas

Real-world AI Factory implementations on Hybrid Manager addressing specific organizational needs and user roles

FAQ

Common questions about deploying, operating, and troubleshooting AI Factory within Hybrid Manager environments

GPU Recommendations

Recommended GPUs and node sizes for the default NIM models used by AI Factory.

Quickstart

Build a simple Gen AI assistant backed by a Knowledge Base and deploy a model endpoint using AI Factory.

Manual

How To Enable AI Factory

Enabling AI Factory in EDB Postgres AI Hybrid Manager

Gen AI Builder

Build intelligent applications using LLMs, knowledge bases, and tools within Hybrid Manager's sovereign infrastructure

Model

Deploy and manage AI models within Hybrid Manager's sovereign infrastructure using Model Library and Model Serving capabilities

Setup

Prerequisites

Environment requirements, GPU planning, and registry configuration needed before deploying AI Factory components

Deploy with HM UI

Step-by-step guide for deploying NVIDIA NIM models and Gen AI applications through the Hybrid Manager web interface

How-to (Use Models)

Using models (Model Clusters)

Accessing KServe endpoints from applications and AIDB; links to hub procedures and examples.

How-to (Use Gen AI builder)

Using models (Builder)

How to consume model endpoints from Gen AI Builder in Hybrid Manager; links to hub how-tos.

Operate & Observe

Observability

Where to observe Gen AI and Model Serving workloads in Hybrid Manager; links to hub.

Troubleshooting

Diagnostic procedures and resolution strategies for common AI Factory issues within Hybrid Manager deployments