Skip to content

dumping tensors with llama-eval-callback #14567

Discussion options

You must be logged in to vote

As far as I am aware, --mmproj does not currently exist as a command-line argument in llama-eval-callback or elsewhere in llama.cpp. The model loader in llama.cpp is primarily designed around text-only models like LLaMA, and while there is emerging support for vision-language models (VLMs), the integration of vision-specific components like mm_proj isn't fully exposed through the CLI yet.

If you're working with a VLM (e.g. Qwen-VL, BLIP-style models), the mm_proj tensor typically corresponds to a linear projection layer used to align visual embeddings with the language model. These weights may exist in the original model files (e.g., in Hugging Face safetensors), but during conversion to .…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@ryan-mangeno
Comment options

Answer selected by ryan-mangeno
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants