β‘ PerformanceBatching BackendOn this pageBatching Backend π Coming Soon: Guide on batching LLM requests for optimal throughput. Previewβ Batching combines multiple requests to reduce overhead: // Coming in v0.5.0val batchedBackend = OpenAIBackend().batched( maxBatchSize = 10, maxWaitTime = 100.milliseconds) See Alsoβ Performance Overview Tool Caching