Replies: 1 comment
-
I assume it can output up to 32K tokens (max_seq_len) - ie. the text it generates for the user, but only have as input (context. maybe that might also include the prompt) up to 8K tokens. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
According to the default configs of the Mistral model, max_seq_len is set to 32k, although the model has only 8k context. Can someone explain to me how max_seq_len differs from context?
Beta Was this translation helpful? Give feedback.
All reactions