How to get response content and timings info in stream mode? #11018
Unanswered
waltersamson
asked this question in
Q&A
Replies: 1 comment
-
The detailed response only available in the last chunk in the stream response. You should test it with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm using llama.cpp server with SillyTavern frontend.
In non-stream mode, a response json, formed by
to_json_non_oaicompat()
, is sent to the frontend, it contains detailed information including response content and timings.However, when I enable stream mode, the response is empty. The verbose log just says:
How can I have the response json sent to the frontend in stream mode? Or at least visible in the logs?
There are sparse information on this in the server README, but those seems outdated.
Is the lack of information in stream mode a limitation, or a bug?
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions