BentoML combines the best developer experience with a focus on operating ML in production.
Our platform enables Data Science teams to do their best work.
Get started with BentoML today!
Copyright © 2022 BentoML
$ pip install bentoml
BentoML's open source model serving framework is blazing fast and easy to get started. Check out the tutorial and documentation for next steps.
An open platform for ML in production
BentoML is compatible across machine learning frameworks and standardizes ML model packaging and management for your team.
Python-first, scales with powerful optimizations
Maximize resource utilization across multi-stage serving pipelines.
Dynamically group predictions requests in real-time into batches for model inference.
Run your model serving workloads seamless with accelerated hardware.
Ship to prod faster
Learn about how we can help accelerate your machine learning projects to production. Save time and resources by streamlining deployment for your development and production workflows.
BentoML integrates with your existing data stack and enable your ML team to continuously deploy and improve their models in production.
Try BENTO CLOUD: The best way to deploy ML