Skip to content
@bentoml

BentoML

The most flexible way to serve AI models in production

Welcome to BentoML šŸ‘‹ Twitter Follow Slack

BentoML

What is BentoML? šŸ‘©ā€šŸ³

BentoML is an open-source model serving library for building performant and scalable AI applications with Python. It comes with everything you need for serving optimization, model packaging, and production deployment.

šŸ”Ø Build Anywhere with Open-Source:

šŸš¢ Efficient scaling on your/our Cloud:

  • ā˜ļø BentoCloud: Inference Platform for enterprise AI teams to build fast, secure, and scalable AI applications.

Get in touch šŸ’¬

šŸ‘‰ Join our Slack community!

šŸ‘€ Follow us on X @bentomlai and LinkedIn

šŸ“– Read our blog

Pinned

  1. BentoML BentoML Public

    The most flexible way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Inference Graph/Pipelines, Compound AI systems, Multi-Modal, RAG as a Service, and more!

    Python 6.5k 738

  2. OpenLLM OpenLLM Public

    Run any open-source LLMs, such as Llama 2, Mistral, as OpenAI compatible API endpoint in the cloud.

    Python 8.8k 554

Repositories

Showing 10 of 70 repositories

Sponsoring

  • @pdm-project

Top languages

Loadingā€¦

Most used topics

Loadingā€¦