More examples¶
Check out the following examples to deploy different inference APIs with BentoML.
BentoVLLM - Accelerate your model inference and improve serving throughput by using vLLM as your LLM backend.
BentoDiffusion - Self-host diffusion models with BentoML to generate custom images and video clips.
BentoXTTS - Convert text to speech based on your custom audio data.
BentoWhisperX - Convert spoken words into text for AI scenarios like virtual assistants, voice-controlled devices, and automated transcription services.
Sentence Transformer - Transform text into numerical vectors for a variety of natural language processing (NLP) tasks.
BentoCLIP - Build a CLIP (Contrastive Language-Image Pre-training) application for tasks like zero-shot learning, image classification, and image-text matching.
BentoBLIP - Leverage BLIP (Bootstrapping Language Image Pre-training) to improve the way AI models understand and process the relationship between images and textual descriptions.
BentoBark - Generate highly realistic audio like music, background noise and simple sound effects with Bark.
BentoYolo - Build an object detection inference API server with YOLO.
BentoChatTTS - Deploy a text-to-speech model ChatTTS for dialogue scenarios like chatbots and virtual assistants.
BentoMoirai - Create a forecasting inference API for time-series data.
BentoResnet - Build an image classification inference API server with ResNet.
BentoLangGraph - Deploy a LangGraph AI agent application with BentoML.
BentoCrewAI - Deploy a CrewAI multi-agent application with BentoML.
See bentoml/examples for more examples.