You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
While it should be possible to configure the batch_size of the Hugging Face pipeline/model under the hood,
this component only accepts a prompt (single str) as input, therefore this practically does not allow batching.
@anakin87 Since the HuggingFaceLocalGenerator only accepts a single str input, I am having to sequentially loop through each prompt. This makes running my code very slow. I would like to perform batched inference so I can make better use of my GPUs and run the inference faster
Is there a way to perform batched inferencing using
HuggingfaceLocalGenerator
? I could not find any information about this on the docs.The text was updated successfully, but these errors were encountered: