Skip to content

Conversation

@joshhan619
Copy link
Contributor

Instead of using HF's PretrainedConfig and dumping all model parameters and hyperparameters into one place, TrainingConfig clearly separates the model's PretrainedConfig from training parameters like batch size, learning rate, etc.

In addition, this PR adds extensions of PretrainedConfig for each baseline (LTSMConfig, DLinearConfig, PatchTSTConfig, and InformerConfig). This exposes the expected model parameters to users and it provides reasonable defaults as backup.

@ynchuang ynchuang merged commit 4cbea68 into datamllab:ltsm-stack Mar 6, 2025
1 check passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants