-
Notifications
You must be signed in to change notification settings - Fork 341
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Getting special characters in model generation #70
Comments
Did you get any solution for this? |
Not yet :( |
I tried with small and medium models it's the same for me.
|
@marianafidalgo ,could you share your output-daily data ? |
Yeah I have encountered the same issues. The model just returns tens of "!!!!!!" and then cannot be conversed with anymore. This behaviour happens after the 4th round of the conversation, like clockwork. The problem seems to step from the implementation of chat history. With the step hardcoded to constant 0, the bot works, albeit without any memory. as the step reaches 3 everything just breaks down. Maybe it's a dataset issue, or maybe it is some sort of memory issue. EDIT: Seems like after a few round the EoS token that should end the round is not longer added after the bot response. |
Did you solve it? |
I did not manage to figure out the root cause of the problem. I did manage to make the bot respond as it should by constraining the lenght of the |
Since it breaks after step 3/4 , a potential hacky solution is maintain a queue with a fixed length of 3 maybe which stores past inputs and outputs and use them rather than the whole history, although some context is lost this would allow the chatbot to run endlessly without breaking down and keeping some context rather than none as in step hardcoded to 0.
When you say EoS is not added, is there a way to add it manually? Like after every response we add EoS , would that fix the issue? |
Changing the number of chat rounds kept in memory proved to solve the issue most of the time, however, it was not as reliable as I needed it to be. As per my response on Sep 17, I have instead taken the length of the tensor into account and using a 'hacky' fix like the one below I was able to get it to work without freaking out at all.
To be absolutely honest I did not pursue this line of thinking since I managed to get it working well enough for my implementation. If adding EoS manually will make it behave properly, I do not know. |
Hello. I fine tuned my DialoGPT small model and trained it on Daily Dialogue database.
When speaking with the model, on its last utterance generation, the outputs has random punctuation characters.
This is my generation:
The output:
The text was updated successfully, but these errors were encountered: