Trained model returning ==Weblinks== #156
Unanswered
robinlaegler
asked this question in
Q&A
Replies: 1 comment 5 replies
-
I don't think batch size is likely the issue. How much data do you have? Increasing would probably help. Also how many epochs did you train for? I have gotten good results around 10k. |
Beta Was this translation helpful? Give feedback.
5 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello, i trained a en-de model using the OPUS-Wikipedia Dataset and ~10.000 sentences of my own training data. I dropped the batch size to 800. I trained it on GoogleColab , because I don't have a GPU laying around. The training finished in about 5 hours. The model is pretty bad and it sometimes returns "==Weblinks==" or "==Einzelnachweise==".
Does anybody dealt with the same problem or knows where these tags are coming from?
And are there any tips to train a better model? Do I need more data or a higher batch size? Which is more important?
Beta Was this translation helpful? Give feedback.
All reactions