-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When using Ollama, a long loop of Initializing Ollama Model happens before any output #443
Comments
@FGhrawi can you pull latest and confirm that this is fixed? |
I did a fresh clone on the latest code and the code keeps wanting to download and run the local_llama. #Set to Use for New OLLAMA provider This is what I have in my defaultCharacter.
Code is also tries to load the wrong OLLAMA embedding model too by default - hence having to set it here. I'm debugging the problem now and working on a fix for it. |
Is this issue still valid?? |
yes, ollama and llama local got merged together and the ollama logic is not working right. getting close to a fix. |
@yodamaster726 what is your character file like? I had ollama as the model in my character and it seemed to use ollama just fine. |
I tried the latest tag v0.1.3 and then tried the latest code from yesterday. My character file was the default one. updates to fix this problem: #521 |
just tested with the latest code and used Seems like its still going in a loop. |
ModelProviderName.LLAMALOCAL is still in loop with alpha.1 tag release. Anyone found a solution? |
Did you try last version we just release ? |
I tried the following fix locally which prevented the looping on LLAMA_LOCAL model provider |
Describe the bug
When using Ollama, the model being used is loaded over and over in a loop (sometimes 10-15+ times) in the terminal before any output or agent decision-making happens.
Fwiw, one model is configured at a time in .env (so no swapping is being done)
To Reproduce
Use any model with the Ollama provider
Expected behavior
Model loads once. (Since ollama is on keepalive and only 1 model is being used)
Screenshots
Additional context
I am on a fresh clone of the repo with a basic character configuration.
Fwiw, I have 24gb vram, and this happens even on smaller models.
The text was updated successfully, but these errors were encountered: