Closed
Description
Originally spotted by @iamlemec in abetlen/llama-cpp-python#1089 reproduced with llama.cpp by passing --no_kv_offload
to ./main
. Bug causes the model to generate repeated #
's instead of a valid completion.
Originally spotted by @iamlemec in abetlen/llama-cpp-python#1089 reproduced with llama.cpp by passing --no_kv_offload
to ./main
. Bug causes the model to generate repeated #
's instead of a valid completion.