dumping tensors with llama-eval-callback #14567
-
I am trying to dump the tensors of a model, its a vlm and I can get the text model tensors but there doesn't seem to be a --mmproj argument and was wondering if this exists elsewhere. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
As far as I am aware, If you're working with a VLM (e.g. Qwen-VL, BLIP-style models), the
To access
Currently, there is no CLI argument equivalent to Would love to work on a feature regarding this, let me know if any help is required. |
Beta Was this translation helpful? Give feedback.
As far as I am aware,
--mmproj
does not currently exist as a command-line argument inllama-eval-callback
or elsewhere inllama.cpp
. The model loader inllama.cpp
is primarily designed around text-only models like LLaMA, and while there is emerging support for vision-language models (VLMs), the integration of vision-specific components likemm_proj
isn't fully exposed through the CLI yet.If you're working with a VLM (e.g. Qwen-VL, BLIP-style models), the
mm_proj
tensor typically corresponds to a linear projection layer used to align visual embeddings with the language model. These weights may exist in the original model files (e.g., in Hugging Facesafetensors
), but during conversion to.…