BentoML is an open-source model serving library for building performant and scalable AI applications with Python. It comes with everything you need for serving optimization, model packaging, and production deployment.
🔨 Build Anywhere with Open-Source:
- 🍱 BentoML: The Unified Model Serving Framework
- 🦾 OpenLLM: Self-hosting Large Language Models Made Easy
🚢 Efficient scaling on your/our Cloud:
- ☁️ BentoCloud: Inference Platform for enterprise AI teams to build fast, secure, and scalable AI applications.
👀 Follow us on X @bentomlai and LinkedIn
📖 Read our blog