Enable LoRA speedup #46
Draft
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What's changing
This improvement makes it possible to finetune whisper large-v3(!) locally or on colab in a 8GB VRAM GPU! The previous code was able to only finetune whisper small in such a machine.
Closes #2
How to test it
Steps to test the changes:
Additional notes for reviewers
The issue is that the model produced by the training is a
Peft Model, meaning you cant load and use it the same way as the standard model from theTrainer. This means that we need to add extra custom code for the Transcription app and the evaluation of the dataset.I already...
/docs)