Closed
Description
Name and Version
version: 4410 (4b0c638)
built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
llama-server
Problem description & steps to reproduce
The /completion
end point use to emit a output.generation_settings.model
with the correct value.
Its now output.model
which is fine, but it seems to always say gpt-3.5-turbo
which is not the model I'm using. I've also tried settings response_fields
to see if that "tricks" it into giving the right value but it doesn't seem to.
When adding the model in the server tool it appears as the attached logs.
First Bad Commit
No response
Relevant log output
llama_model_loader: - kv 0: general.architecture str = llama
llama_model_loader: - kv 1: general.type str = model
llama_model_loader: - kv 2: general.name str = Hermes 3 Llama 3.1 8B
llama_model_loader: - kv 3: general.organization str = NousResearch
llama_model_loader: - kv 4: general.basename str = Hermes-3-Llama-3.1