Open
Description
Hi @kashif, thanks for your great work in implementing all these Transformer models! I noticed that for many models, especially the long sequence time series forecasting models (Auto/ETS/NS-Transformer), you have decided to remove the enc/dec_embedding
layers for the dynamic, real inputs and directly set d_model = self.input_size * len(self.lags_seq) + self._number_of_features
as inputs to the Transformer layer (please correct me if I got this wrong). This makes the hyperparameter d_model
not tunable, but tied to the inputs. Could I ask what prompted this decision?
Metadata
Metadata
Assignees
Labels
No labels