More examples¶

Check out the following examples to deploy different inference APIs with BentoML.

  • BentoVLLM - Accelerate your model inference and improve serving throughput by using vLLM as your LLM backend.

  • BentoDiffusion - Self-host diffusion models with BentoML to generate custom images and video clips.

  • BentoXTTS - Convert text to speech based on your custom audio data.

  • BentoWhisperX - Convert spoken words into text for AI scenarios like virtual assistants, voice-controlled devices, and automated transcription services.

  • Sentence Transformer - Transform text into numerical vectors for a variety of natural language processing (NLP) tasks.

  • BentoCLIP - Build a CLIP (Contrastive Language-Image Pre-training) application for tasks like zero-shot learning, image classification, and image-text matching.

  • BentoBLIP - Leverage BLIP (Bootstrapping Language Image Pre-training) to improve the way AI models understand and process the relationship between images and textual descriptions.

  • BentoBark - Generate highly realistic audio like music, background noise and simple sound effects with Bark.

  • BentoYolo - Build an object detection inference API server with YOLO.

  • BentoChatTTS - Deploy a text-to-speech model ChatTTS for dialogue scenarios like chatbots and virtual assistants.

  • BentoMoirai - Create a forecasting inference API for time-series data.

  • BentoResnet - Build an image classification inference API server with ResNet.

  • BentoLangGraph - Deploy a LangGraph AI agent application with BentoML.

  • BentoCrewAI - Deploy a CrewAI multi-agent application with BentoML.

See bentoml/examples for more examples.