Add bisenetv2. My implementation of BiSeNet
-
Updated
Feb 5, 2023 - Python
Add bisenetv2. My implementation of BiSeNet
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
This repository deploys YOLOv4 as an optimized TensorRT engine to Triton Inference Server
The Triton backend for the ONNX Runtime.
ClearML - Model-Serving Orchestration and Repository Solution
Deploy stable diffusion model with onnx/tenorrt + tritonserver
Deploy DL/ ML inference pipelines with minimal extra code.
OpenAI compatible API for TensorRT LLM triton backend
Hardware-accelerated DNN model inference ROS 2 packages using NVIDIA Triton/TensorRT for both Jetson and x86_64 with CUDA-capable GPU
An Alternative for Triton Inference Server. Boosting DL Service Throughput 1.5-4x by Ensemble Pipeline Serving with Concurrent CUDA Streams for PyTorch/LibTorch Frontend and TensorRT/CVCUDA, etc., Backends
This repository is an AI bootcamp material that consist of a workflow for computer vision
Compare multiple optimization methods on triton to imporve model service performance
Build Recommender System with PyTorch + Redis + Elasticsearch + Feast + Triton + Flask. Vector Recall, DeepFM Ranking and Web Application.
Set up CI in DL/ cuda/ cudnn/ TensorRT/ onnx2trt/ onnxruntime/ onnxsim/ Pytorch/ Triton-Inference-Server/ Bazel/ Tesseract/ PaddleOCR/ NVIDIA-docker/ minIO/ Supervisord on AGX or PC from scratch.
Advanced inference pipeline using NVIDIA Triton Inference Server for CRAFT Text detection (Pytorch), included converter from Pytorch -> ONNX -> TensorRT, Inference pipelines (TensorRT, Triton server - multi-format). Supported model format for Triton inference: TensorRT engine, Torchscript, ONNX
Provides an ensemble model to deploy a YoloV8 ONNX model to Triton
Diffusion Model for Voice Conversion
Triton face detection & recognition
Magface Triton Inferece Server Using Tensorrt
Add a description, image, and links to the triton-inference-server topic page so that developers can more easily learn about it.
To associate your repository with the triton-inference-server topic, visit your repo's landing page and select "manage topics."