• This repository has been archived on 26/Feb/2024
  • Stars
    star
    172
  • Rank 216,573 (Top 5 %)
  • Language
    Python
  • License
    Other
  • Created over 2 years ago
  • Updated 4 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Fast model deployment on any cloud 🚀

🚀 Fast model deployment on any cloud

actions_status docs join_slack

bentoctl helps deploy any machine learning models as production-ready API endpoints on the cloud, supporting AWS SageMaker, AWS Lambda, EC2, Google Compute Engine, Azure, Heroku and more.

👉 Join our Slack community today!

✨ Looking deploy your ML service quickly? You can checkout BentoML Cloud for the easiest and fastest way to deploy your bento. It's a full featured, serverless environment with a model repository and built in monitoring and logging.

Highlights

  • Framework-agnostic model deployment for Tensorflow, PyTorch, XGBoost, Scikit-Learn, ONNX, and many more via BentoML: the unified model serving framework.
  • Simplify the deployment lifecycle of deploy, update, delete, and rollback.
  • Take full advantage of BentoML's performance optimizations and cloud platform features out-of-the-box.
  • Tailor bentoctl to your DevOps needs by customizing deployment operator and Terraform templates.

Getting Started

Supported Platforms:

Upcoming

Custom Operator

Users can built custom bentoctl plugin from the deployment operator template to deploy to cloud platforms not yet supported or to internal infrastructure.

If you are looking for deploying with Kubernetes, check out Yatai: Model deployment at scale on Kubernetes.

Installation

pip install bentoctl

| 💡 bentoctl designed to work with BentoML version 1.0.0 and above. For BentoML 0.13 or below, you can use the pre-v1.0 branch in the operator repositories and follow the instruction in the README. You can also check out the quickstart guide for 0.13 here.

Community

Contributing

There are many ways to contribute to the project:

  • Create and share new operators. Use deployment operator template to get started.
  • If you have any feedback on the project, share it with the community in Github Discussions under the BentoML repo.
  • Report issues you're facing and "Thumbs up" on issues and feature requests that are relevant to you.
  • Investigate bugs and reviewing other developer's pull requests.

Usage Reporting

BentoML and bentoctl collects usage data that helps our team to improve the product. Only bentoctl's CLI commands calls are being reported. We strip out as much potentially sensitive information as possible, and we will never collect user code, model data, model names, or stack traces. Here's the code for usage tracking. You can opt-out of usage tracking by setting environment variable BENTOML_DO_NOT_TRACK=True:

export BENTOML_DO_NOT_TRACK=True

Licence

Elastic License 2.0 (ELv2)

More Repositories

1

OpenLLM

Run any open-source LLMs, such as Llama 2, Mistral, as OpenAI compatible API endpoint in the cloud.
Python
9,124
star
2

BentoML

The easiest way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Multi-model Inference Graph/Pipelines, LLM/RAG apps, and more!
Python
6,714
star
3

Yatai

Model Deployment at Scale on Kubernetes đŸĻ„ī¸
TypeScript
771
star
4

OneDiffusion

OneDiffusion: Run any Stable Diffusion models and fine-tuned weights with ease
Python
325
star
5

stable-diffusion-server

Deploy Your Own Stable Diffusion Service
Python
191
star
6

gallery

BentoML Example Projects 🎨
Python
134
star
7

OCR-as-a-Service

Turn any OCR models into online inference API endpoint 🚀 🌖
Python
47
star
8

transformers-nlp-service

Online Inference API for NLP Transformer models - summarization, text classification, sentiment analysis and more
Python
41
star
9

CLIP-API-service

CLIP as a service - Embed image and sentences, object recognition, visual reasoning, image classification and reverse image search
Jupyter Notebook
36
star
10

BentoVLLM

Self-host LLMs with vLLM and BentoML
Python
32
star
11

simple_di

Simple dependency injection framework for Python
Python
19
star
12

yatai-deployment

🚀 Launching Bento in a Kubernetes cluster
Go
16
star
13

Fraud-Detection-Model-Serving

Online model serving with Fraud Detection model trained with XGBoost on IEEE-CIS dataset
Jupyter Notebook
14
star
14

aws-sagemaker-deploy

Fast model deployment on AWS Sagemaker
Python
14
star
15

yatai-image-builder

đŸŗ Build OCI images for Bentos in k8s
Go
14
star
16

sentence-embedding-bento

Sentence Embedding as a Service
Jupyter Notebook
14
star
17

google-cloud-run-deploy

Fast model deployment on Google Cloud Run
Python
13
star
18

aws-lambda-deploy

Fast model deployment on AWS Lambda
Python
13
star
19

aws-ec2-deploy

Fast model deployment on AWS EC2
Python
13
star
20

IF-multi-GPUs-demo

Python
13
star
21

rag-tutorials

a series of tutorials implementing rag service with BentoML and LlamaIndex
Python
11
star
22

diffusers-examples

API serving for your diffusers models
Python
10
star
23

BentoSVD

Python
9
star
24

Pneumonia-Detection-Demo

Pneumonia Detection - Healthcare Imaging Application built with BentoML and fine-tuned Vision Transformer (ViT) model
Python
8
star
25

yatai-chart

Helm Chart for installing Yatai on Kubernetes ⎈
Mustache
7
star
26

benchmark

BentoML Performance Benchmark 🆚
Jupyter Notebook
7
star
27

plugins

the swish knife to all things bentoml.
Starlark
6
star
28

bentoctl-operator-template

Python
6
star
29

heroku-deploy

Deploy BentoML bundled models to Heroku
Python
6
star
30

BentoLMDeploy

Self-host LLMs with LMDeploy and BentoML
Python
5
star
31

bentoml-core

Rust
5
star
32

BentoControlNet

Python
4
star
33

BentoWhisperX

Python
4
star
34

google-compute-engine-deploy

HCL
4
star
35

BentoCLIP

building a CLIP application using BentoML
Python
4
star
36

BentoRAG

Tutorial: Build RAG Apps with Custom Models Served with BentoML
Python
4
star
37

quickstart

BentoML Quickstart Example
Python
4
star
38

deploy-bento-action

A GitHub Action to deploy bento to cloud
3
star
39

azure-functions-deploy

Fast model deployment on Azure Functions
Python
3
star
40

azure-container-instances-deploy

Fast model deployment on Azure container instances
Python
3
star
41

containerize-push-action

docker's build-and-push-action equivalent for bentoml
TypeScript
3
star
42

BentoSentenceTransformers

how to build a sentence embedding application using BentoML
Python
2
star
43

BentoTRTLLM

Python
2
star
44

bentoml-arize-fraud-detection-workshop

Jupyter Notebook
2
star
45

BentoSDXLTurbo

how to build an image generation application using BentoML
Python
2
star
46

yatai-schemas

Go
1
star
47

bentoctl-workshops

Python
1
star
48

llm-bench

Python
1
star
49

bentocloud-homepage-news

1
star
50

yatai-common

Go
1
star
51

BentoBLIP

how to build an image captioning application on top of a BLIP model with BentoML
Python
1
star
52

BentoYolo

BentoML service of YOLO v8
Python
1
star
53

.github

✨🍱đŸĻ„ī¸
1
star
54

BentoBark

Python
1
star
55

BentoMLCLLM

Python
1
star
56

BentoTGI

Python
1
star
57

openllm-benchmark

Python
1
star