Quantized versions of Qwen3-VL-30B-A3B-Instruct with mmproj

my motivation for this was that most/all other quants i seen of this is they only have an .mmproj file for f16,
but this file needs to be of same quant as the model itself
(locking you to only using f16 if you want the vision support, which is ofcourse the whole point of the model)

NOTE: im inexperienced in quantizing, its possible i have done mistakes or missed something, possibly reducing or degrading the quality of this model

Downloads last month
366
GGUF
Model size
31B params
Architecture
qwen3vlmoe
Hardware compatibility
Log In to view the estimation

2-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for EnderCrypt/Qwen3-VL-30B-A3B-Instruct-GGUF-mmproj

Quantized
(42)
this model