generated from amazon-archives/__template_Custom
-
Notifications
You must be signed in to change notification settings - Fork 183
Open
Labels
enhancementNew feature or requestNew feature or request
Description
Is your feature request related to a problem?
TextSimilarityCrossEncoderModel predictions are performed one by one inside a loop using getPredictor().predict(input)
rather than processing the inputs in batches using batchPredict(listOfInputs)
This approach looks inefficient. Is there a specific reason for processing the predictions one by one?
To my understanding, Predictor.java creates internally a list/batch with a single element and calls batchPredict anyway.
Similar behaviour is presented on TextEmbeddingModel
What solution would you like?
Use batchProcessing
to improve efficiency
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or request
Type
Projects
Status
In Progress