You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
then, checked by a OpenWebUI dev who says may be an issue on ollama-end...
Set context length then read logfile. these are 2 logs examples showing my context length had an extra zero added
time=2024-05-08T20:12:24.868Z level=WARN source=memory.go:17 msg="requested context length is greater than model max context length" requested=81920 model=65536
time=2024-05-08T05:24:12.168Z level=WARN source=memory.go:17 msg="requested context length is greater than model max context length" requested=20480 model=8192
also
appears there may be a 2nd (related?) issue/bug with Max Tokens
when set to 2048... in a conversation, getting an extra zero getting added here too?... see below two examples of 20480 which should have been 2048
...
time=2024-05-09T14:33:27.120Z level=WARN source=server.go:77 msg="requested context length is greater than the model's training context window size" requested=20480 "training size"=4096
...
.................................................................................................
llama_new_context_with_model: n_ctx = 20480
OS
Linux
GPU
Nvidia
CPU
No response
Ollama version
0.1.34
The text was updated successfully, but these errors were encountered:
What is the issue?
first reported to OpenWebUI at: open-webui/open-webui#2141
then, checked by a OpenWebUI dev who says may be an issue on ollama-end...
Set context length then read logfile. these are 2 logs examples showing my context length had an extra zero added
time=2024-05-08T20:12:24.868Z level=WARN source=memory.go:17 msg="requested context length is greater than model max context length" requested=81920 model=65536
time=2024-05-08T05:24:12.168Z level=WARN source=memory.go:17 msg="requested context length is greater than model max context length" requested=20480 model=8192
also
appears there may be a 2nd (related?) issue/bug with Max Tokens
when set to 2048... in a conversation, getting an extra zero getting added here too?... see below two examples of 20480 which should have been 2048
...
time=2024-05-09T14:33:27.120Z level=WARN source=server.go:77 msg="requested context length is greater than the model's training context window size" requested=20480 "training size"=4096
...
.................................................................................................
llama_new_context_with_model: n_ctx = 20480
OS
Linux
GPU
Nvidia
CPU
No response
Ollama version
0.1.34
The text was updated successfully, but these errors were encountered: