You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have been trying to adapt the prompt template. It kind of works, even without changing the template, but eventually it goes into an endless response loop. I guess that the issue are the stop tokens or something like that.
Has anyone tried this?
The text was updated successfully, but these errors were encountered:
Quite sure you are right, the reason will most probably be the different prompt formats ("chat templates") the models use. This project is a bit outdated and therefore still uses the raw template format.
Meanwhile most of this can be abstracted away. Today most LLM providers offer servers with chat endpoints, like these:
Alternatively you can also use then python OpenAI library to wrap the base endpoints (http://localhost:1234/v1). For example code, Linguflex project does some LLM requests in this way.
Thanks for this great project. It works quite well.
However, I would like to change the model to:
https://huggingface.co/mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated-GGUF
I have been trying to adapt the prompt template. It kind of works, even without changing the template, but eventually it goes into an endless response loop. I guess that the issue are the stop tokens or something like that.
Has anyone tried this?
The text was updated successfully, but these errors were encountered: