A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.
-
Updated
May 29, 2024 - Python
A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.
Triton backend for https://github.com/OpenNMT/CTranslate2
Package for running Nvidia Triton within python test with features like Dockerfile DSL and building images on fly.
Learnings and experimentation with GPU programming
🐳 Scripps Whale Acoustics Lab 🌎 Scripps Acoustic Ecology Lab - Triton with remoras in development
Efficient kernel for RMS normalization with fused operations, includes both forward and backward passes, compatibility with PyTorch.
ClearML - Model-Serving Orchestration and Repository Solution
Deploy DL/ ML inference pipelines with minimal extra code.
Hardware-accelerated, deep learned model support for object detection including YOLOv8 and DetectNet
LAMB go brrr
Binary Ninja plugin that can be used to apply Triton's dead store eliminitation pass on basic blocks or functions.
MLModelService wrapping Nvidia's Triton Server
This bootcamp is designed to give NLP researchers an end-to-end overview on the fundamentals of NVIDIA NeMo framework, complete solution for building large language models. It will also have hands-on exercises complimented by tutorials, code snippets, and presentations to help researchers kick-start with NeMo LLM Service and Guardrails.
Flash linear attention kernels in Triton
Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackable.
삼각형의 실전! Triton
Add a description, image, and links to the triton topic page so that developers can more easily learn about it.
To associate your repository with the triton topic, visit your repo's landing page and select "manage topics."