NVIDIA Launches Inference Platforms for Large Language Models and Generative AI Workloads
Each platform is optimized for in-demand workloads, including AI video, image generation, large language model deployment and recommender inference.
- Each platform is optimized for in-demand workloads, including AI video, image generation, large language model deployment and recommender inference.
- “The rise of generative AI is requiring more powerful inference computing platforms,” said Jensen Huang, founder and CEO of NVIDIA.
- With NVIDIA AI Enterprise, customers receive NVIDIA Enterprise Support, regular security reviews and API stability for NVIDIA Triton Inference Server, TensorRT and more than 50 pretrained models and frameworks.
- Hands-on labs for trying the NVIDIA inference platform for generative AI are available immediately at no cost on NVIDIA LaunchPad .