Model Serving Made EasyΒΆ

pypi status Downloads Actions Status Documentation Status join BentoML Slack

BentoML is a flexible, high-performance framework for serving, managing, and deploying machine learning models.

  • Supports Multiple ML frameworks, including Tensorflow, PyTorch, Keras, XGBoost and more

  • Cloud native deployment with Docker, Kubernetes, AWS, Azure and many more

  • High-Performance online API serving and offline batch serving

  • Web dashboards and APIs for model registry and deployment management

BentoML bridges the gap between Data Science and DevOps. By providing a standard interface for describing a prediction service, BentoML abstracts away how to run model inference efficiently and how model serving workloads can integrate with cloud infrastructures. See how it works!

πŸ’» Get started with BentoML: Quickstart Guide | Quickstart on Google Colab

πŸ‘©β€πŸ’» Star/Watch/Fork the BentoML Github Repository.

πŸ‘‰ Join the community Slack and discussions on Github.