Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 825 162

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 434 75

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.9k 1.7k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.8k 246

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4.3k 511

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.9k 1k

Repositories

Showing 10 of 710 repositories
  • NemoClaw Public

    Run OpenClaw more securely inside NVIDIA OpenShell with managed inference

    NVIDIA/NemoClaw’s past year of commit activity
    TypeScript 19,496 Apache-2.0 2,421 227 (1 issue needs help) 176 Updated Apr 20, 2026
  • warp Public

    A Python framework for GPU-accelerated simulation, robotics, and machine learning.

    NVIDIA/warp’s past year of commit activity
    Python 6,533 Apache-2.0 486 193 12 Updated Apr 20, 2026
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 16,099 3,850 352 (1 issue needs help) 352 Updated Apr 20, 2026
  • topograph Public

    A toolkit for discovering cluster network topology.

    NVIDIA/topograph’s past year of commit activity
    Go 116 Apache-2.0 20 8 4 Updated Apr 20, 2026
  • spark-rapids-jni Public

    RAPIDS Accelerator JNI For Apache Spark

    NVIDIA/spark-rapids-jni’s past year of commit activity
    Cuda 56 Apache-2.0 80 88 11 Updated Apr 20, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 13,423 2,300 593 741 Updated Apr 20, 2026
  • cuda-python Public

    CUDA Python: Performance meets Productivity

    NVIDIA/cuda-python’s past year of commit activity
    Cython 3,223 274 198 35 Updated Apr 20, 2026
  • ncx-infra-controller-core Public

    NCX Infra Controller - Hardware Lifecycle Management and multitenant networking

    NVIDIA/ncx-infra-controller-core’s past year of commit activity
    Rust 125 Apache-2.0 79 160 (5 issues need help) 60 Updated Apr 20, 2026
  • cuda-quantum Public

    C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows

    NVIDIA/cuda-quantum’s past year of commit activity
    C++ 1,005 Apache-2.0 363 440 (16 issues need help) 122 Updated Apr 20, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 2,516 Apache-2.0 362 56 129 Updated Apr 20, 2026