BentoML is an open-source model serving library for building performant and scalable AI applications with Python. It comes with everything you need for serving optimization, model packaging, and production deployment.
šØ Build Anywhere with Open-Source:
- š± BentoML: The Unified Model Serving Framework
- š¦¾ OpenLLM: Self-hosting Large Language Models Made Easy
š¢ Efficient scaling on your/our Cloud:
- āļø BentoCloud: Inference Platform for enterprise AI teams to build fast, secure, and scalable AI applications.
š Join our Slack community!
š Follow us on X @bentomlai and LinkedIn
š Read our blog