Generative AI

NVIDIA AI Foundry

The end-to-end platform and service for building

# custom models for generative AI.

[Get Started](https://www.nvidia.com/en-us/ai.md)

[Press Release](https://nvidianews.nvidia.com/news/nvidia-ai-foundry-custom-llama-generative-models) | [Blog](https://blogs.nvidia.com/blog/ai-foundry-enterprise-generative-ai)

Overview

## Generative AI Built by You, for You

NVIDIA AI Foundry is a platform and service for building custom generative AI models with enterprise data and domain-specific knowledge. Just as TSMC manufactures chips designed by other companies, NVIDIA AI Foundry enables organizations to develop their own AI models.

A chip foundry provides state-of-the-art transistor technology, manufacturing process, large chip fabs, expertise, and a rich ecosystem of third-party tools and library providers. Similarly, NVIDIA AI Foundry includes NVIDIA-created AI models like Nemotron and Edify, popular open [foundation models](https://build.nvidia.com/explore/discover), [NVIDIA NeMo™](https://www.nvidia.com/en-us/ai-data-science/generative-ai/nemo-framework.md) software for customizing models, and dedicated capacity on [NVIDIA DGX™ Cloud](https://www.nvidia.com/en-us/data-center/dgx-cloud.md)—built and backed by NVIDIA AI experts. The output is [NVIDIA NIM™](http://ai.nvidia.com)—an inference microservice that includes the custom model, optimized engines, and a standard API—which can be deployed anywhere.

NVIDIA AI Foundry and its libraries are integrated into the world’s leading AI ecosystem of startups, enterprise software providers, and global service providers.

### Customize Generative AI Models for Enterprise Applications With Llama 3.1

With NVIDIA AI Foundry, enterprises can build generative AI solutions tailored for their specific organizational, industrial, or cultural needs.

[Read Now](https://developer.nvidia.com/blog/customize-generative-ai-models-for-enterprise-applications-with-llama-3-1/)

## How Does NVIDIA AI Foundry Work?

The NVIDIA AI Foundry uses enterprise data, along with synthetically generated data, to augment and alter the general knowledge contained in a pretrained foundation model. Once the model is customized, evaluated, and has guardrails, it’s output as an NVIDIA NIM inference microservice. Developers use the NIM’s standard API to build generative AI-powered applications. Knowledge gained from applications in deployment can be fed back into the foundry to further improve custom models.

Benefits

## Explore the Benefits of NVIDIA AI Foundry

### Models Built to Your Needs

Start with state-of-the-art foundation models and then create custom models built specifically for your applications’ needs, using your enterprise data and domain expertise.

### Faster Time to Solution

Focus on harnessing the knowledge of your organization and leveraging the game-changing insights of AI, instead of maintaining and tuning your AI development platform.

### Production Ready

Protect data privacy and intellectual property by creating your own models. Run confidently in production environments with NVIDIA Enterprise Support, API stability, and reliable security updates.

### Unprecedented Performance

Build models efficiently using the latest NVIDIA architecture on DGX Cloud with your preferred cloud service providers (CSPs). Deploy custom models with optimized engines, packaged as an easy-to-use NVIDIA NIM inference microservice.

## Components

### Build Custom Models for Generative AI Enterprise Applications

#### Start With Optimized Foundation Models

Generative AI Models

Enterprises can start with community and NVIDIA-built models, optimized to deliver the best performance on NVIDIA accelerated infrastructure. These models can be customized with NVIDIA NeMo and deployed in production at scale anywhere with NVIDIA NIM.

Start customizing optimized [large language](https://build.nvidia.com/explore/reasoning), [vision language](https://build.nvidia.com/explore/vision), [speech](https://build.nvidia.com/explore/speech), and [healthcare](https://build.nvidia.com/explore/healthcare) models to achieve higher accuracy for your use cases.

[Read the Blog](https://developer.nvidia.com/blog/search-posts/?q=foundation+models)

[Explore All Models](https://build.nvidia.com/explore/discover)

#### Fine-Tune Foundation Models With Enterprise Data

NVIDIA NeMo

NVIDIA NeMo is an end-to-end platform for developing custom generative AI—including large language models (LLMs), multimodal, vision, and speech AI—anywhere. Deliver enterprise-ready models with precise data curation, cutting-edge customization, and optimized performance.

It includes NeMo Curator for accelerated data curation, NeMo Customizer for simplified fine-tuning, NeMo Evaluator for evaluating AI models, and NeMo Guardrails for safeguarding your LLMs.

[Read the Blogs](https://developer.nvidia.com/blog/tag/nemo/)

[Learn More About NeMo](https://www.nvidia.com/en-us/ai-data-science/generative-ai/nemo-framework.md)

#### Build Models With the Latest NVIDIA Architecture

NVIDIA DGX Cloud

To customize an AI model requires access to significant computing resources, AI expertise, and software to streamline AI model development.

NVIDIA DGX Cloud is an end-to-end AI platform for developers, offering dedicated capacity built on the latest NVIDIA AI architecture and co-engineered with the world’s leading CSPs. Build your custom AI models on DGX Cloud.

[Read the Blog](https://developer.nvidia.com/blog/tag/dgx-cloud/)

[Learn More About DGX Cloud](https://www.nvidia.com/en-us/data-center/dgx-cloud.md)

#### Run Models in Production Anywhere

NVIDIA NIM™

The output of the NVIDIA AI Foundry is an NVIDIA NIM to quickly deploy and scale the custom model anywhere.

NVIDIA NIM is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI model inferencing across clouds, data centers, and workstations.

[Read the Blog](https://developer.nvidia.com/blog/nvidia-nim-offers-optimized-inference-microservices-for-deploying-ai-models-at-scale/)

[Learn More About NIM](https://www.nvidia.com/en-us/ai.md)

#### Integrate Into Leading Enterprise Platforms

NVIDIA AI Ecosystem

Hundreds of  NVIDIA technology partners are integrating NVIDIA NIM, part of NVIDIA AI Enterprise, into their platforms to speed generative AI deployments for domain-specific applications.

Partners include the major public clouds, system builders, enterprise infrastructure providers, MLOps and AIOps leaders, and many others.

[See the ecosystem](#ecosystem)

#### Develop With Experts by Your Side

NVIDIA Partners

NVIDIA AI Foundry is built and backed by NVIDIA AI experts.

NVIDIA and the NVIDIA Partner Network (NPN)—including leading global systems integrators, service delivery partners, and solution providers—are ready to help you build custom models for your enterprise applications.

[Explore NVIDIA Partners](https://www.nvidia.com/en-us/about-nvidia/partners.md)

Use Cases

## How NVIDIA AI Foundry Is Being Used

See how NVIDIA AI Foundry supports industry use cases and jump-starts your AI development.

1. **AI Chatbot**
2. **Content Generation**
3. **Network Operations Centers**
4. **Intelligent Document Processing**
5. **Security Vulnerability Analysis**
6. **Hyperpersonalized Shopping**

### AI Chatbots

Organizations are looking to build smarter AI chatbots using custom LLMs and retrieval-augmented generation (RAG). With RAG, chatbots can accurately answer domain-specific questions by retrieving current information from an organization’s knowledge base and providing real-time responses in natural language. These chatbots can be used to enhance customer support, personalize AI avatars, manage enterprise knowledge, streamline employee onboarding, provide intelligent IT support, create content, and more.

[Learn More About AI Chatbots](https://www.nvidia.com/en-us/use-cases/ai-for-customer-support.md)

### Content Generation

Generative AI makes it possible to generate highly relevant, bespoke, and accurate content grounded in the domain expertise and proprietary IP of your enterprise.

[Learn More About Content Generation](https://www.nvidia.com/en-us/use-cases/content-creation-using-generative-ai.md)

Katana

### Network Operations Centers

Global telecommunications companies are exploring how to cost-effectively deliver new AI applications to the edge over 5G and upcoming 6G networks. With NVIDIA accelerated computing and AI, telcos, CSPs, and enterprises can build high-performance cloud-native networks—both fixed and wireless—with improved energy efficiency and security.

[Learn More About Generative AI for Network Operation Centers](https://www.nvidia.com/en-us/use-cases/network-operations-assist.md)

Agility, Apptronik, Fourier Intelligence, Unitree

### Intelligent Document Processing (IDP)

In financial services, processing documents involves complex data, such as loan records, external regulatory filings, transaction records, public market filings, and more. Financial institutions can use custom generative AI for IDP, such as building chatbots with RAG to automate loan processes or developing market insights for portfolio construction and trade execution.

[Learn More About Intelligent Document Processing](https://www.nvidia.com/en-us/use-cases/intelligent-document-processing.md)

### Security Vulnerability Analysis

Patching software security issues is becoming progressively more challenging as the number of reported security flaws in the common vulnerabilities and exposures (CVE) database hit a record high in 2022. Generative AI can improve vulnerability defense while decreasing the load on security teams.

[Learn More About Security Vulnerability Analysis](https://www.nvidia.com/en-us/ai-data-science/ai-workflows/security-vulnerability-analysis.md)

Delta Electronics

### AI for Hyperpersonalized Shopping

A retailer’s inventory is complex, with thousands if not millions of products that change seasonally. In the immensely competitive marketplace that’s emerged over the past decade, leading retailers have harnessed the power of AI and data science to offer real-time, hyperpersonalized customer experiences that increase cart size, build brand affinity, and increase conversion.

[Learn More About Hyperpersonalized Shopping](https://www.nvidia.com/en-us/use-cases/hyper-personalized-shopping.md)

Verneek

### Starting Options

## Ways to Get Started With NVIDIA AI Foundry

Use the right tools and technologies to build custom generative AI models.

### Try

Test and build with a wide range of foundation models on ai.nvidia.com.

[Explore Models](https://build.nvidia.com/explore/discover)

### Customize

Use NVIDIA NeMo software and the low-rank adaptation (LoRA) customization technique for LLMs to build a custom model.

[Start Customizing](https://www.nvidia.com/en-us/ai-data-science/generative-ai/nemo-framework.md)

### Fine-Tune

Need dedicated access to the latest NVIDIA AI technology? Train your custom AI model using DGX Cloud, and scale with ease as your AI demands change.

[Get Accelerated Computing](https://www.nvidia.com/en-us/data-center/dgx-cloud.md)

Customer Stories

## How Industry Leaders Are Driving Innovation With NVIDIA AI Foundry

Enterprise Software

### Bringing Intelligent Workflow Automation With Custom LLMs

**Customer**: ServiceNow

**Products**: NVIDIA AI Foundry, NVIDIA NeMo, NVIDIA DGX Cloud

**Technologies**: LLMs

[Read Case Study](https://nvidianews.nvidia.com/news/servicenow-and-nvidia-announce-partnership-to-build-generative-ai-across-enterprise-it)

Amdocs

Telecommunications

### Amdocs Accelerates Generative AI Performance with NVIDIA

**Customer**: Amdocs

**Products**: NVIDIA AI Foundry, NVIDIA NeMo

**Technologies**: LLMs

[Read Case Study](https://developer.nvidia.com/blog/amdocs-accelerates-generative-ai-performance-and-lowers-costs-with-nvidia-nim/)

Media and Entertainment

### Picture This: Getty Images Releases Generative AI by iStock

**Customer**: Getty

**Products**: NVIDIA AI Foundry, NVIDIA Picasso

**Technologies**: Vision language models

[Read Case Study](https://blogs.nvidia.com/blog/nvidia-picasso-istock-generative-ai/)

Ecosystem

## Leading Enterprise Platforms

Partners include the major public clouds, system builders, enterprise infrastructure providers, MLOps and AIOps leaders, and many others.

1. **Model Builders**
2. **AIOps and MLOps Partners**
3. **DGX Cloud Providers**
4. **DGX Cloud Partners**
5. **Production Deployment Partners**
6. **Consulting Services Partners**

Next Steps

## Ready to Get Started?

Talk to an NVIDIA expert about taking custom generative AI applications from pilot to production with the security, API stability, and support of NVIDIA AI Enterprise.

[Contact Sales](https://www.nvidia.com/en-us/data-center/products/ai-enterprise/contact-sales.md)