Pytorch domain library for recommendation systems
-
Updated
Jun 11, 2024 - Python
CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.
Pytorch domain library for recommendation systems
Main repository for QMCPACK, an open-source production level many-body ab initio Quantum Monte Carlo code for computing the electronic structure of atoms, molecules, and solids with full performance portable GPU support
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes
A model-independent chemistry module for atmosphere models
Open Voice OS Status Page
FlashInfer: Kernel Library for LLM Serving
A high-throughput and memory-efficient inference and serving engine for LLMs
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
(in progress) SAH kd-tree parallel construction algorithm implementation
Containers for machine learning
The PennyLane-Lightning plugin provides a fast state-vector simulator written in C++ for use with PennyLane
The open-source serverless GPU container runtime.
Implementations of various simulations for integrate and fire models, as well as conductance based models with synaptic neurotransmission
A retargetable MLIR-based machine learning compiler and runtime toolkit.
Created by Nvidia
Released June 23, 2007