LibreChat always enforces a max context length of 4096 #6894
Replies: 2 comments
-
A quick update: I've added Also tried to modify custom:
- name: Ollama
(...)
addParams:
num_ctx: 16384 ... but I still get the 4095 max context length message as shown above. |
Beta Was this translation helpful? Give feedback.
-
Seems like the only way is to set up the context length in the UI, as described by @danny-avila here: #4259 (comment) In the same thread he also mentions that in the future this will be a config file setting. Closing this as my problem is solved, hope this helps someone in the future. |
Beta Was this translation helpful? Give feedback.
-
I'm using the dockerized LibreChat, Ollama backend, with the cogito:32b model (supports up to 128k context length).
When I try to send a message with more than 4096 tokens, I get the following error message on the LibreChat web UI:
Looking at the docker logs, I also see:
I've tried editing
librechat.yaml
and adding the following:But I still get the same error after restarting all the docker containers (with
docker compose ...
).Using Ollama on the command line, the model seems to accept at least 32k tokens without a problem.
Am I doing something wrong here? What is the right way to increase the context length?
Beta Was this translation helpful? Give feedback.
All reactions