-
Notifications
You must be signed in to change notification settings - Fork 9.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to do conversation with the llama-2-7B-chat model. #846
Comments
We don't have a full chat program example in the repo, but you can adapt the example to build one. Take a look at this thread for a related conversation: #162 |
@Harsh-raj You can use LangChain's ConversationalRetrievalChain example or ConversationChain with ConversationBufferMemory example. |
I am now able to do conversation with the llama-2-7b-chat model. But when max prompt length exceeds the max sequence length the conversation abruptly terminates. I wanted to remove the oldest context of the conversation from the model's memory and make space for the next user prompt. Is this possible? |
Also for running inference of the llama-2-7b-chat model somehow, |
Are you using ConversationChain or ConversationalRetrievalChain to do conversation? You can then remove earlier Q/A pairs in the |
Hey, hope you doing well. I am able to run inference on the llama-2-7B-chat model successfully with the example python script provided. I am new to working and experimenting with large language models. I wanted to know how can i do conversation with the model where the model will consider its previous user prompts chat completion context too for answering next user prompt. I am currently experimenting with the dialogue list present in the example python script but it seems that i will have to go through all of the code and make changes in it. Any guidance is much appreciated. Thank you!
The text was updated successfully, but these errors were encountered: