Qwen3-4B quantized to 4-bit with bitsandbytes. Original model is here.

Note: This model is for being used with diffusers, this repository doesn't contain the tokenizer

Downloads last month
73
Safetensors
Model size
4B params
Tensor type
F32
BF16
U8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support