AI Factory in Hybrid Manager v1.3
The AI Factory workload in Hybrid Manager brings scalable AI, machine learning, and Gen AI capabilities to your Hybrid Manager platform. It enables you to operationalize AI across your Hybrid Manager–managed clusters and data — with deep integration across Postgres, vector search, model serving, and AI assistants.
With AI Factory in Hybrid Manager, you can:
- Deploy Gen AI assistants and agents for internal or external-facing use
- Serve AI models at scale with integrated KServe-powered inferencing and GPU acceleration
- Create Knowledge Bases and perform Retrieval-Augmented Generation (RAG) with vector search
- Manage your model library and deploy trusted models under Hybrid Manager governance
- Integrate AI features directly into your applications and data pipelines
Hub quick links: AI Factory Hub — Gen AI (hub) — Model Serving
For in-depth architecture and core concepts, see the AI Factory Hub.
Example AI solutions you can build
AI Factory in Hybrid Manager enables solutions across many domains:
Enterprise search and knowledge assistants Build RAG-based assistants that integrate with corporate documents, databases, and intranet content.
Customer support chatbots Deploy assistants powered by your own data and domain-specific models, combining conversational AI with Knowledge Bases.
AI-driven data apps Expose AI-powered endpoints for apps — semantic search, recommendations, similarity search, or natural language querying.
Operational AI for internal tools Build AI agents and tools to assist with DevOps, customer success, HR automation, sales enablement, and more.
Domain-specific model serving Serve proprietary or fine-tuned models (LLMs, embedding models, ranking models) as scalable inference services within your Hybrid Manager cluster.
You can start small — with a single assistant or model endpoint — and scale to full AI-powered applications using the AI Factory architecture.
Learning paths
Follow structured learning paths to master AI Factory in Hybrid Manager:
- AI Factory 101 — Introductory concepts and usage
- AI Factory 201 — Building and managing Gen AI and AI Factory workloads
- AI Factory 301 — Advanced integration, scaling, governance, and optimization
Use cases and solutions
We provide detailed guidance and patterns to help you build solutions with AI Factory:
- Common Use Cases — Proven patterns for AI Factory–powered applications
- Industry Solutions — Industry-specific examples and best practices
AI Factory in Hybrid Manager workloads
Hybrid Manager supports a full set of AI Factory capabilities, tightly integrated into its control plane:
Gen AI workloads
- Gen AI in Hybrid Manager — Gen AI capabilities in Hybrid Manager
- Agent Studio & Builder (consolidated) — Create assistants, tools, structures, rulesets; manage Knowledge Bases
Model management and serving
- Model Library — Manage and govern your model assets
- Model Serving — Deploy and scale model inference services on Kubernetes
- GPU resource management — Configure and allocate GPU capacity for serving
Vector Engine
- Vector Engine — Integrated vector search and similarity capabilities with Postgres
Hybrid Manager learn content
In addition to AI Factory content in the hub, Hybrid Manager provides additional Learn content for Hybrid Manager–specific usage:
How AI Factory fits within Hybrid Manager
Hybrid Manager is the control plane for your databases, analytics, and AI workloads. AI Factory complements other HM capabilities:
- Databases: Provision and manage Postgres clusters; AI Factory connects to these for vector search, RAG, and application endpoints.
- Analytics: Offload data to Iceberg/Delta and query via SQL; AI Factory’s Pipelines and Vector Engine integrate with these datasets for retrieval.
- Operations: Use HM observability and governance to monitor Gen AI, Pipelines, and model serving alongside databases and analytics.
Prerequisites for AI Factory in HM:
- A Hybrid Manager project with AI Factory enabled and required entitlements for Gen AI, Pipelines, Vector Engine, and Model Serving.
- Access to GPU resources where needed for serving. See GPU resource management.
Get started
To begin building with AI Factory in Hybrid Manager:
- Follow the AI Factory learning paths
- Explore use cases and industry solutions
- Read How-To Guides
- Deploy your first Gen AI assistant or model using Gen AI in HM or Model Serving
AI Factory in Hybrid Manager gives you a scalable, secure platform to operationalize AI across your hybrid data estate.
Getting started
Getting started
Quick orientation to AI Factory on Hybrid Manager with pointers to hub concepts and how-tos.
Architecture
Technical architecture of AI Factory components within Hybrid Manager's Kubernetes infrastructure
Sovereign AI
Applying Sovereign AI principles with AI Factory in HM.
Use Cases & Personas
Real-world AI Factory implementations on Hybrid Manager addressing specific organizational needs and user roles
FAQ
Common questions about deploying, operating, and troubleshooting AI Factory within Hybrid Manager environments
GPU Recommendations
Recommended GPUs and node sizes for the default NIM models used by AI Factory.
Quickstart
Build a simple Gen AI assistant backed by a Knowledge Base and deploy a model endpoint using AI Factory.
Manual
How To Enable AI Factory
Enabling AI Factory in EDB Postgres AI Hybrid Manager
Gen AI Builder
Build intelligent applications using LLMs, knowledge bases, and tools within Hybrid Manager's sovereign infrastructure
Model
Deploy and manage AI models within Hybrid Manager's sovereign infrastructure using Model Library and Model Serving capabilities
Setup
Prerequisites
Environment requirements, GPU planning, and registry configuration needed before deploying AI Factory components
Deploy with HM UI
Step-by-step guide for deploying NVIDIA NIM models and Gen AI applications through the Hybrid Manager web interface
How-to (Use Models)
Using models (Model Clusters)
Accessing KServe endpoints from applications and AIDB; links to hub procedures and examples.
How-to (Use Gen AI builder)
Using models (Builder)
How to consume model endpoints from Gen AI Builder in Hybrid Manager; links to hub how-tos.
Operate & Observe
Observability
Where to observe Gen AI and Model Serving workloads in Hybrid Manager; links to hub.
Troubleshooting
Diagnostic procedures and resolution strategies for common AI Factory issues within Hybrid Manager deployments