A universal scalable machine learning model deployment solution
-
Updated
Jun 6, 2024 - Java
A universal scalable machine learning model deployment solution
Large Language Model Text Generation Inference
📚 Jupyter notebook tutorials for OpenVINO™
A high-throughput and memory-efficient inference and serving engine for LLMs
Espero que en este repo encuentres inspiración para aprender y desarrollarte en el mundo de la Estadística, no soy perfecto en todo así que si tienes una sugerencia la aceptares con todo el gusto, espero disfrutes lo que puedes encontrar en la página **Este repo aun esta en construcción**
Utilities to use the Hugging Face Hub API
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Faster Whisper transcription with CTranslate2
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
A fast, easy-to-use, production-ready inference server for computer vision supporting deployment of many popular model architectures and fine-tuned models.
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
Port of OpenAI's Whisper model in C/C++
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"
TypeDB: the polymorphic database powered by types
Python package to perform statistical hypothesis tests.
Making large AI models cheaper, faster and more accessible
带你从零实现一个高性能的深度学习推理库,支持大模型 llama2 、Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library step by step
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Add a description, image, and links to the inference topic page so that developers can more easily learn about it.
To associate your repository with the inference topic, visit your repo's landing page and select "manage topics."