Batch inference
Also known as: offline inference
Processing large queued inputs where latency per item matters less than total cost - common for document extraction or nightly scoring.
Also known as: offline inference
Processing large queued inputs where latency per item matters less than total cost - common for document extraction or nightly scoring.
Contact if you need a term added for a security or procurement review.