Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 384 66

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 354 50

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.1k 1.5k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.6k 216

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 3.6k 395

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.4k 813

Repositories

Showing 10 of 590 repositories
  • KAI-Scheduler Public

    KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale

    NVIDIA/KAI-Scheduler’s past year of commit activity
    Go 770 Apache-2.0 83 27 (2 issues need help) 21 Updated Aug 24, 2025
  • cuda-quantum Public

    C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows

    NVIDIA/cuda-quantum’s past year of commit activity
    C++ 779 273 385 (19 issues need help) 78 Updated Aug 24, 2025
  • k8s-dra-driver-gpu Public

    NVIDIA DRA Driver for GPUs

    NVIDIA/k8s-dra-driver-gpu’s past year of commit activity
    Go 416 Apache-2.0 86 53 27 Updated Aug 24, 2025
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 13,326 3,033 284 122 Updated Aug 24, 2025
  • JAX-Toolbox Public

    JAX-Toolbox

    NVIDIA/JAX-Toolbox’s past year of commit activity
    Python 330 Apache-2.0 63 79 41 Updated Aug 24, 2025
  • TensorRT-LLM Public

    TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    C++ 11,433 Apache-2.0 1,702 718 374 Updated Aug 24, 2025
  • NV-Kernels Public

    Ubuntu kernels which are optimized for NVIDIA server systems

    NVIDIA/NV-Kernels’s past year of commit activity
    C 57 41 0 4 Updated Aug 24, 2025
  • cccl Public

    CUDA Core Compute Libraries

    NVIDIA/cccl’s past year of commit activity
    C++ 1,878 260 1,023 (6 issues need help) 158 Updated Aug 24, 2025
  • k8s-nim-operator Public

    An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.

    NVIDIA/k8s-nim-operator’s past year of commit activity
    Go 124 Apache-2.0 31 4 24 Updated Aug 24, 2025
  • k8s-test-infra Public

    K8s-test-infra

    NVIDIA/k8s-test-infra’s past year of commit activity
    Go 9 Apache-2.0 7 0 7 Updated Aug 24, 2025