-
Notifications
You must be signed in to change notification settings - Fork 304
FAQ
Run aichat
with the environment variable AICHAT_LOG_LEVEL=debug
. Then check the log file at <aichat-config-dir>/aichat.log
.
The log contains runtime details and API request and response data, which can help you troubleshoot.
Sessions require context memory. However, The chat API is stateless. Therefore, each request needs to include the chat history. Chat history grows rapidly with the conversation, leading to a rapid increase in data transmitted to the API server.
This presents two problems:
- The amount of data the LLM needs to process for each request keeps increasing, resulting in longer response times and increased costs.
- The data volume may exceed the LLM's processing capacity, leading to errors.
AICHat's strategy is to automatically compress the chat history when the number of tokens in the chat history exceeds a certain value (compress_threshold
).
There are several possible reasons for this, which need to be investigated one by one:
1. The LLM doesn't support stream function calling, only non-stream function calling.
SOLUTION: Add -S
CLI option or Run .set stream false
REPL command to switch to non-stream mode.
2. The <aichat-config-dir>/functions/functions.json
file is missing or empty.
SOLUTION: Enter the llm-functions
directory and rebuild the tools/agents.
3: The input text is not related to the functions.
SOLUTION: Adjust the input text to be relevant to the functions.
Roles act as a sort of prompt library, and agents are similar to OpenAI's GPTs.
An agent is a superset of a role; a role is equivalent to an agent with only the instructions
field, while agents have additional features: variables, conversation starters, documents (RAG), and AI tools.