Unified Model Serving Framework#

github_stars pypi_status actions_status documentation_status join_slack

BentoML is an open-source model serving library for building performant and scalable AI applications with Python. It comes with everything you need for serving optimization, model packaging, and production deployment.

Start your BentoML journey#

The BentoML documentation provides detailed guidance on the project with hands-on tutorials and examples. If you are a first-time user of BentoML, we recommend that you read the following documents in order:

Gain a basic understanding of the BentoML open-source framework, its workflow, installation, and a quickstart example.

Create different BentoML projects for common machine learning scenarios, like large language models, image generation, embeddings, speech recognition, and more.

Dive into BentoML’s features and advanced use cases, including GPU support, clients, monitoring, and performance optimization.

A fully managed platform for deploying and scaling BentoML in the cloud.

Stay informed#

The BentoML team uses the following channels to announce important updates like major product releases and share tutorials, case studies, as well as community news.

To receive release notifications, star and watch the BentoML project on GitHub. For release notes and detailed changelogs, see the Releases page.