Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
57 commits
Select commit Hold shift + click to select a range
fc6d2ee
velocity encoder
WojciechMat Oct 25, 2023
d43008a
fixed average dist, training for single token
WojciechMat Oct 25, 2023
c2ab363
lower resolution start in quantization
WojciechMat Oct 25, 2023
410d03f
run on cuda:0 by default
WojciechMat Oct 25, 2023
9a4e19b
new masked midi encoder for T5 unsupervised learning
WojciechMat Oct 25, 2023
82a6331
clean
WojciechMat Oct 25, 2023
ea29489
comment
WojciechMat Oct 25, 2023
6d7934a
cleaner better code, MaskedMidiDataset
WojciechMat Oct 25, 2023
b944ebc
denoising training, remove lr schedule
WojciechMat Oct 26, 2023
f0e4601
training on a large dataset
WojciechMat Oct 27, 2023
8a0c5ee
fix dataset building
WojciechMat Oct 27, 2023
5becd1d
optimization - use np and torch
WojciechMat Oct 27, 2023
f8dd742
fix AT record slicing...
WojciechMat Oct 27, 2023
89c9459
fix bug in masking
WojciechMat Oct 28, 2023
ba00168
remove printing
WojciechMat Oct 28, 2023
69f3104
filter incorrect sequences
WojciechMat Oct 28, 2023
5752447
denoise checkpoint dir
WojciechMat Oct 29, 2023
02cfcc2
denosing initial dashboard
WojciechMat Oct 29, 2023
4c29459
fix untokenize, add decode method to maskedmidiencoder
WojciechMat Oct 29, 2023
333f179
print tokens on a dashboard
WojciechMat Oct 31, 2023
c7251f4
decoder_start_id = start_id
WojciechMat Nov 1, 2023
1d89012
fix:
WojciechMat Nov 1, 2023
3ad7214
MaskedNoteEncoder
WojciechMat Nov 6, 2023
f4d44b8
fix dashboard
WojciechMat Nov 6, 2023
570f5c0
T5 dstart config
WojciechMat Nov 1, 2023
36b2968
apply quantization
WojciechMat Nov 6, 2023
731a0ef
fix cls token hiding from loss
WojciechMat Nov 6, 2023
3033a56
eos token = pad token, PEP 526 variable annotations
WojciechMat Nov 8, 2023
80f3717
download_model and masking
WojciechMat Nov 9, 2023
51e8f5f
change config
WojciechMat Nov 10, 2023
8fffb9c
pre-training
WojciechMat Nov 10, 2023
3358d6b
vocab size fix
WojciechMat Nov 10, 2023
6c92834
fix vocab size
WojciechMat Nov 11, 2023
5363b9d
remove prints
WojciechMat Nov 11, 2023
b289bfe
change default model path
WojciechMat Nov 11, 2023
87899a9
finetuned velocity model
WojciechMat Nov 12, 2023
29d1ec6
clean
WojciechMat Nov 12, 2023
192eeca
change model name, change run name for finetuning
WojciechMat Nov 13, 2023
079d020
change distance calculation
WojciechMat Nov 14, 2023
bbf94bf
legacy
WojciechMat Nov 15, 2023
9fcb65e
update dashboard
WojciechMat Nov 15, 2023
681dd0c
fix finetuning
WojciechMat Nov 15, 2023
ea22e1a
expand source on click
WojciechMat Nov 17, 2023
8bfacef
merge
WojciechMat Nov 16, 2023
f7b68a2
dashboard for denoising on single-token-per-note tokenization
WojciechMat Nov 20, 2023
ef8567a
fix base_lr when finetuning
WojciechMat Nov 22, 2023
e19d90e
cleanup, configs modification
WojciechMat Dec 6, 2023
e755672
fix dashboard, pre_defined_architectures added
WojciechMat Dec 8, 2023
becf0f9
denoise checkpoint
WojciechMat Dec 26, 2023
0bcf484
streamlit_pianoroll
WojciechMat Nov 13, 2023
011f4c9
change distance calculation
WojciechMat Nov 14, 2023
96444f4
better dashboard
WojciechMat Nov 14, 2023
d083ab6
validation every 1000*log_frequency steps
WojciechMat Nov 14, 2023
96379d7
change config
WojciechMat Nov 14, 2023
965d337
200*log_frequency
WojciechMat Nov 14, 2023
d6dd095
merge
WojciechMat Nov 15, 2023
5f4b377
pre-commit hooks
WojciechMat Dec 26, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Empty file added checkpoints/denoise/.keep
Empty file.
46 changes: 46 additions & 0 deletions configs/T5denoise-dstart.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,46 @@
train:
num_epochs: 5
accum_iter: 5
batch_size: 2
base_lr: 3e-5
warmup: 4000
finetune: False

model_name: T5
dataset_name: 'roszcz/maestro-v1-sustain'
target: denoise
seed: 26

overfit: False

tokens_per_note: single
time_quantization_method: dstart
masking_probability: 0.2
mask: tokens

encoder: velocity
time_bins: 100

dataset:
sequence_len: 128
sequence_step: 42

quantization:
dstart: 5
duration: 5
velocity: 3

device: "cuda:0"

log: True
log_frequency: 10
run_name: midi-T5-${now:%Y-%m-%d-%H-%M}
project: "midi-hf-transformer"

pre_defined_model: null
model:
d_model: 512
d_kv: 64
d_ff: 2048
num_layers: 6
num_heads: 8
47 changes: 47 additions & 0 deletions configs/T5denoise.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,47 @@
train:
num_epochs: 5
accum_iter: 10
batch_size: 8
base_lr: 3e-5
warmup: 4000
finetune: False

model_name: T5
dataset_name: 'roszcz/maestro-v1-sustain'
target: denoise
seed: 26

overfit: False

tokens_per_note: multiple
time_quantization_method: start
masking_probability: 0.15
mask: notes

encoder: velocity
time_bins: 100

dataset:
sequence_duration: 5
sequence_step: 2

quantization:
start: 50
duration: 5
velocity: 3

device: "cuda:0"

log: True
log_frequency: 10
run_name: midi-T5-${now:%Y-%m-%d-%H-%M}
project: "midi-hf-transformer"

pre_defined_model: null

model:
d_model: 512
d_kv: 64
d_ff: 2048
num_layers: 6
num_heads: 8
5 changes: 4 additions & 1 deletion configs/T5start.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@ seed: 26

overfit: False

tokens_per_note: multiple
time_quantization_method: start
dataset:
sequence_duration: 5
Expand All @@ -23,13 +24,15 @@ dataset:
duration: 3
velocity: 3

device: "cpu"
device: "cuda:0"

log: True
log_frequency: 10
run_name: midi-T5-${now:%Y-%m-%d-%H-%M}
project: "midi-hf-transformer"

pre_defined_model: null

model:
d_model: 512
d_kv: 64
Expand Down
45 changes: 45 additions & 0 deletions configs/T5velocity-dstart.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
train:
num_epochs: 5
accum_iter: 10
batch_size: 8
base_lr: 3e-5
finetune: True
warmup: 4000


pretrained_checkpoint: midi-T5-2023-11-15-17-18.pt
model_name: T5
dataset_name: 'roszcz/maestro-v1-sustain'
target: velocity
seed: 26
time_bins: 100

overfit: False

tokens_per_note: "single"
time_quantization_method: dstart
dataset:
sequence_len: 128
sequence_step: 42

quantization:
dstart: 5
duration: 5
velocity: 3

device: "cuda:0"

log: True
log_frequency: 10
run_name: midi-T5-${now:%Y-%m-%d-%H-%M}
project: "midi-hf-transformer"

pre_defined_model: null

model:
d_model: 512
d_kv: 64
d_ff: 2048
num_layers: 6
num_decoder_layers: None
num_heads: 8
9 changes: 7 additions & 2 deletions configs/T5velocity.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,31 +4,36 @@ train:
batch_size: 8
base_lr: 3e-4
warmup: 4000
finetune: True

pretrained_checkpoint: midi-T5-2023-11-11-10-29.pt
model_name: T5
dataset_name: 'roszcz/maestro-v1-sustain'
target: velocity
seed: 26

overfit: False

tokens_per_note: "multiple"
time_quantization_method: start
dataset:
sequence_duration: 5
sequence_step: 2

quantization:
start: 400
start: 20
duration: 3
velocity: 3

device: "cpu"
device: "cuda:0"

log: True
log_frequency: 10
run_name: midi-T5-${now:%Y-%m-%d-%H-%M}
project: "midi-hf-transformer"

pre_defined_model: null

model:
d_model: 512
d_kv: 64
Expand Down
5 changes: 5 additions & 0 deletions configs/architectures/large.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
d_model: 512
d_kv: 64
d_ff: 2048
num_layers: 6
num_heads: 8
5 changes: 5 additions & 0 deletions configs/architectures/mid.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
d_model: 256
d_kv: 32
d_ff: 1024
num_layers: 6
num_heads: 8
5 changes: 5 additions & 0 deletions configs/architectures/small.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
d_model: 256
d_kv: 32
d_ff: 512
num_layers: 4
num_heads: 4
Loading