Skip to content

Misc. bug: Server response "model" seems to be incorrect #11069

Closed
@turtle0x1

Description

@turtle0x1

Name and Version

version: 4410 (4b0c638)
built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu

Operating systems

Linux

Which llama.cpp modules do you know to be affected?

llama-server

Problem description & steps to reproduce

The /completion end point use to emit a output.generation_settings.model with the correct value.

Its now output.model which is fine, but it seems to always say gpt-3.5-turbo which is not the model I'm using. I've also tried settings response_fields to see if that "tricks" it into giving the right value but it doesn't seem to.

When adding the model in the server tool it appears as the attached logs.

First Bad Commit

No response

Relevant log output

llama_model_loader: - kv   0:                       general.architecture str              = llama
llama_model_loader: - kv   1:                               general.type str              = model
llama_model_loader: - kv   2:                               general.name str              = Hermes 3 Llama 3.1 8B
llama_model_loader: - kv   3:                       general.organization str              = NousResearch
llama_model_loader: - kv   4:                           general.basename str              = Hermes-3-Llama-3.1

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions