Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 723 129

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 414 66

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.7k 1.6k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.8k 237

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4.1k 481

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.8k 986

Repositories

Showing 10 of 679 repositories
  • numbast Public

    Numbast is a tool to build an automated pipeline that converts CUDA APIs into Numba bindings.

    NVIDIA/numbast’s past year of commit activity
    Python 57 Apache-2.0 18 28 (3 issues need help) 9 Updated Feb 28, 2026
  • KAI-Scheduler Public

    KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale

    NVIDIA/KAI-Scheduler’s past year of commit activity
    Go 1,150 Apache-2.0 156 25 (1 issue needs help) 73 Updated Feb 28, 2026
  • warp Public

    A Python framework for accelerated simulation, data generation and spatial computing.

    NVIDIA/warp’s past year of commit activity
    Python 6,254 Apache-2.0 446 184 7 Updated Feb 28, 2026
  • nvidia-resiliency-ext Public

    NVIDIA Resiliency Extension is a python package for framework developers and users to implement fault-tolerant features. It improves the effective training time by minimizing the downtime due to failures and interruptions.

    NVIDIA/nvidia-resiliency-ext’s past year of commit activity
    Python 264 45 2 17 Updated Feb 28, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 12,970 2,134 536 563 Updated Feb 28, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 2,071 Apache-2.0 284 68 97 Updated Feb 28, 2026
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 15,461 3,639 301 (1 issue needs help) 316 Updated Feb 28, 2026
  • pants Public Forked from pantsbuild/pants

    The Pants Build System

    NVIDIA/pants’s past year of commit activity
    Python 4 Apache-2.0 716 0 0 Updated Feb 28, 2026
  • nvshmem Public

    NVIDIA NVSHMEM is a parallel programming interface for NVIDIA GPUs based on OpenSHMEM. NVSHMEM can significantly reduce multi-process communication and coordination overheads by allowing programmers to perform one-sided communication from within CUDA kernels and on CUDA streams.

    NVIDIA/nvshmem’s past year of commit activity
    C++ 469 62 18 18 Updated Feb 28, 2026
  • gpu-driver-container Public

    The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.

    NVIDIA/gpu-driver-container’s past year of commit activity
    Shell 160 Apache-2.0 76 23 35 Updated Feb 28, 2026