Skip to content

Support for ensemble inference #996

Answered by parano
elia-secchi asked this question in General
Discussion options

You must be logged in to vote

Hi @Rasen-wq

Great question! I have considered this use case when initially designing BentoML. Besides running computations in parallel, it is also useful for users that need to fetch additional data from feature store or third-party APIs (e.g. a fraud detection service that needs to fetch credit score from another service provider while other pre-processing computation can run in parallel).

We did prototype a version with Celery and it worked really well, although it was before we introduced micro-batching and I assume that branch no longer works. I'm not sure if Celery is still the best choice here but I think we do want to expose an API for users to define multiple steps that can run i…

Replies: 2 comments 6 replies

Comment options

You must be logged in to vote
4 replies
@yubozhao
Comment options

@elia-secchi
Comment options

@yubozhao
Comment options

@elia-secchi
Comment options

Comment options

You must be logged in to vote
2 replies
@elia-secchi
Comment options

@danield137
Comment options

Answer selected by elia-secchi
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
4 participants