GitHub - NVIDIA/nim-deploy: A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment. (original) (raw)
Introduction
This repo showcases different ways NVIDIA NIMs can be deployed. This repo contains reference implementations, example documents, and architecture guides that can be used as a starting point to deploy multiple NIMs and other NVIDIA microservices into Kubernetes and other production deployment environments.
NoteThe content in this repository is designed to provide reference architectures and best-practices for production-grade deployments and product integrations; however the code is not validated on all platforms and does not come with any level of enterprise support. While the deployments should perform well, please treat this codebase as experimental and a collaborative sandbox. For long-term production deployments that require enterprise support from NVIDIA, looks to the official releases on NVIDIA NGC which are based on the code in this repo.
Deployment Options
Category | Deployment Option | Description |
---|---|---|
On-premise Deployments | Helm | |
LLM NIM | ||
Open Source Platforms | ||
KServe | ||
Independent Software Vendors | ||
Cloud Service Provider Deployments | Azure | |
AKS Managed Kubernetes | ||
Azure ML | ||
Azure prompt flow | ||
Amazon Web Services | ||
EKS Managed Kubernetes | ||
Amazon SageMaker | ||
EKS Managed Kubernetes - NIM Operator | ||
Google Cloud Platform | ||
GKE Managed Kubernetes | ||
Google Cloud Vertex AI | ||
Cloud Run | ||
NVIDIA DGX Cloud | ||
NVIDIA Cloud Functions | ||
Documents | Deployment Guide | |
Hugging Face NIM Deployment |
Contributions
Contributions are welcome. Developers can contribute by opening a pull request and agreeing to the terms in CONTRIBUTING.MD.
Support and Getting Help
Please open an issue on the GitHub project for any questions. All feedback is appreciated, issues, requested features, and new deployment scenarios included.