Quantized versions of Qwen3-VL-30B-A3B-Instruct with mmproj
my motivation for this was that most/all other quants i seen of this is they only have an .mmproj file for f16,
but this file needs to be of same quant as the model itself
(locking you to only using f16 if you want the vision support, which is ofcourse the whole point of the model)
NOTE: im inexperienced in quantizing, its possible i have done mistakes or missed something, possibly reducing or degrading the quality of this model
- Downloads last month
- 366
Hardware compatibility
Log In
to view the estimation
2-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for EnderCrypt/Qwen3-VL-30B-A3B-Instruct-GGUF-mmproj
Base model
Qwen/Qwen3-VL-30B-A3B-Instruct