Replies: 1 comment 1 reply
-
I think so, starting from a pre-trained weights should be better, as the model see more data, pre-training tasks such as supervised or self-supervised approaches can help the model converge faster on downstream task and achieve better performance. But be aware of the domain gaps, if the pre-training data and fine-tuning data are unrelevent, there could be more challenges for the model. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Dear MONAI Label team,
thank you for the great project! I have been using MONAI Label to develop my own custom interactive segmentation models, similar to DeepEdit and DeepGrow, and I was wondering about something very simple, but not much discussed in the literature.
Does non-interactive pre-training help when training interactive segmentation models? For example, if we train DeepEdit for 90 epochs without any clicks (just zero out the guidance maps) and then fine-tune for 10 epochs with 10 clicks per iteration would that give similar results as training with 10 clicks per iteration for 100 epochs? I can imagine that the model might learn the correlation between its errors and the positions of new clicks during the first epochs but have not seen any experiments regarding this.
I would be happy to hear about your experience with this since training from scratch for every new interactive configuration takes a lot of time and is not very scalable/sustainable. I saw that the original DeepEdit paper considered using 0%, 25%, and 50% click-free iterations, but I have not seen a mix of, e.g., 100% click-free for pre-training and then 0% click-free for fine-tuning.
Best,
Zdravko
Beta Was this translation helpful? Give feedback.
All reactions