#AIGlossary
batch inference
#GoogleCloud
The process of inferring predictions on multiple unlabeled examples divided into smaller subsets ("batches").
Batch inference can take advantage of the parallelization features of accelerator chips. That is, multiple accelerators can simultaneously infer predictions on different batches of unlabeled examples, dramatically increasing the number of inferences per second.
https://developers.google.com/machine-learning/glossary#batch-inference
