Change the repository type filter
All
Repositories list
582 repositories
- C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows
- Ongoing research training transformer models at scale
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
- CUDA Core Compute Libraries
- BioNeMo Framework: For building and adapting AI models in drug discovery at scale
- TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.
- AIStore: scalable storage for AI applications
- Differentiable signal processing on the sphere for PyTorch
- LLM KV cache compression made easy
JAX-Toolbox
Public- NVIDIA Federated Learning Application Runtime Environment
- Examples for Recommenders - easy to train and deploy on accelerated infrastructure.
- A Python framework for accelerated simulation, data generation and spatial computing.