Analog Model Training Taking Longer in AIHWKIT, is this Normal? #695
-
|
Hi, • My RPU configuration seems correct for both loading and HWA training. Also, is it always necessary to start with a digital model before converting to analog, or can I train directly in analog from the start? What would you recommend? Thank you for your suggestions! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
|
Can you share the configuration you used for the Analog HWA. This is not a normal behavior. It might be an issue for the way you configured the experiment. Please look at the example we used: https://github.com/IBM/aihwkit/blob/master/examples/06_lenet5_hardware_aware.py |
Beta Was this translation helpful? Give feedback.
Thanks for your insights!
I have resolved the issue with longer training epochs in my analog model using AIHWKIT. For small networks, is it necessary to use pretrained digital weights before converting to analog? Does it primarily serve as weight initialization, or is there a reason we can not train the entire model directly in analog from the start?
Would skipping the digital pretraining significantly impact the models performance or training efficiency, especially in handling non-idealities? What is generally recommended for smaller networks?