Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LoRA for Whisper speech transcription #483

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

adhulipa
Copy link

Description

Based off of a large chunk of work from the LLM LoRA example, this PR presents applying LoRA fine-tuning to the Whisper speech model. All of the relevant changes to run LoRA for Whisper are in a new directory called lora on {project-root}/whisper/lora. The primary training & data-loading scripts are lora.py, utils.py. The LoRA layer definitions are implemented in whisper/lora/models/lora.pyand mimic the existing work for LLM LoRA.

Core changes for Whisper are:

  1. Adapting the train() func to batch up audio & transcriptions pairs as inputs
  2. Applying LoRA layers to both audio encoder and text decoder blocks of the whisper model

All other changes are essentially ancillary changes to keep the whisper-lora example self-contained and easy to run and modify.

  1. Included 3-4 VS Code run configurations to Train, Fuse, and Transcribe using the new lora-adapated whisper models
  2. Some new helper scripts such as whisper/run_transcribe.py
  3. Duplicated the existing whisper inference code aa a sub-folder under the new {project-root}/whisper/lora/models directory. This was primarily because I didn't want to re-use the existing whisper code ({project-root}/whisper) with relative path imports etc in my new lora code ({project-root}/whisper/lora). It seemed like doing so would work fine in some run configurations but may not at other times. To keep things simple and easily hackable/flexible I duplicated the existing whisper modeling code as-is into {project-root}/whisper/lora/models. Therefore, almost the entirety of {project-root}/whisper/lora/models should be identical to existing{project-root}/whisper/.

Tests/Runs

  1. Download whisper mlx models from hugging face
huggingface-cli download mlx-community/whisper-medium-mlx --local-dir /path/to/projects/mlx-examples/whisper/mlx_models/whisper-medium-mlx --local-dir-use-symlinks False
  1. Transcribe using this model on audio file from some language (ex: Telugu)
↪ /some/python /path/to/projects/mlx-examples/whisper/run_transcribe.py --audio /path/to/projects/mlx-examples/whisper/whisper/assets/adityateluguthursday.wav --model /path/to/projects/mlx-examples/whisper/mlx_models/whisper-medium-mlx 
 বেবববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববব�
  1. Train on speech dataset, such as mozilla-foundation/common_voice_16_1 or its variants
↪ /some/python /path/to/projects/mlx-examples/whisper/lora/lora.py --model  /path/to/projects/mlx-examples/whisper/mlx_models/whisper-medium-mlx --train --adapter-file  /path/to/projects/mlx-examples/whisper/lora_adapters_whisper_with_telugu.npz --hf-dataset mozilla-foundation/common_voice_16_1 --hf-dataset-lang te --batch-size 1 --lora-layers 4 
Loading pretrained model  /path/to/projects/mlx-examples/whisper/mlx_models/whisper-medium-mlx
Applying LoRA parameters to AudioEncoder...
Done applying Encoder LoRA Linear layers
Encoder: Total parameters 305.811M
Encoder: Trainable parameters 0.131M
Applying LoRA parameters to TextDecoder...
Done applying Decoder LoRA Linear layers
Decoder: Total parameters 456.773M
Decoder: Trainable parameters 0.131M
Finished adding LoRA params! :)
Loading datasets
Using hf dataset mozilla-foundation/common_voice_16_1...
Using dataset lang te...
Loading dataset mozilla-foundation/common_voice_16_1, te from hugging face
Loaded datasets with 38 train, 25 valid, 30 test
Training
Iter 1: Val loss 2256.857, Val took 10.683s
Iter 10: Train loss 2276.202, It/sec 1.816, 
Iter 20: Train loss 1937.338, It/sec 1.600, 
Iter 30: Train loss 2230.634, It/sec 1.487, 
Iter 40: Train loss 1349.048, It/sec 1.614, 
Iter 50: Train loss 1162.343, It/sec 1.582, 
Iter 60: Train loss 912.588, It/sec 1.590, 
Iter 70: Train loss 592.943, It/sec 1.664, 
Iter 80: Train loss 962.219, It/sec 1.624, 
Iter 90: Train loss 712.701, It/sec 1.682, 
Iter 100: Train loss 692.088, It/sec 1.677, 
Iter 100: Saved adapter weights to /path/to/projects/mlx-examples/whisper/lora_adapters_whisper_with_telugu.npz.
Iter 110: Train loss 613.315, It/sec 1.646, 
.....
  1. Fuse trained adapters.npz with base model
↪ /some/python /path/to/projects/mlx-examples/whisper/lora/fuse.py --model  /path/to/projects/mlx-examples/whisper/mlx_models/whisper-medium-mlx --adapter-file /path/to/projects/mlx-examples/whisper/lora_adapters_whisper_with_telugu.npz --save-path /path/to/projects/mlx-examples/whisper/lora_fused_model_whisper_with_telugu 
  1. Transcribe using the new fused model
↪ /some/python /path/to/projects/mlx-examples/whisper/run_transcribe.py --audio /path/to/projects/mlx-examples/whisper/whisper/assets/adityateluguthursday.wav --model /path/to/projects/mlx-examples/whisper/lora_fused_model_whisper_with_telug 
 I have to go to office on weekdays

The telugu language phrase I used in my docs here roughly means “i have to go to the office on thursday”. But the model seems to have transcribed it to “I have to go to office on weekdays.“, which is a pretty good translation.
I didn’t intend to train a translator but i was kinda impressed that it did so. I should note: this quirky and delightful instance happened in one of my training runs; all the other times the transcription wasn't great. It's not super reproducible because I only ran the training for ~1000 iterations on ~38 or so training examples. I trained this on a m1max MacBook Pro with 64gb of memory for about ~10-12mins. The converging val loss I saw was ~50-70 after 1000 iterations.

Some other instances of training runs results in

↪ /some/python /path/to/projects/mlx-examples/whisper/run_transcribe.py --audio /path/to/projects/mlx-examples/whisper/whisper/assets/adityateluguthursday.wav --model /path/to/projects/mlx-examples/whisper/lora_fused_model_whisper_with_telugu
 Good morning. Good morning. Good ... .

or

.....whisper/lora_fused_model_whisper_with_telugu
 I I I

or

....whisper/lora_fused_model_whisper_with_telugu 
 Namastar. My name is Adithya. I am very happy to meet you. Namastar.

which in all cases is definitely better than transcribing বেবববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববববব� which is the output of the original model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

1 participant