Minimax M2.1 gguf
Make sure you have enough memory/GPU.
You may try prompts from a dataset MiniMaxAI-VIBE to vibe code (meaning Minimax M2.1 generating applications for you -- ios, android, web, desktop applications).
Use the model in ollama
- First download and install ollama.
Note: the official ollama models do not have Qwen3-Next support yet. You need do the following.
- Command in windows command line (or mac os, linux), or in terminal in ubuntu, type:
ollama run hf.co/John1604/MiniMax-M2.1-gguf:q3_k_m
(q3_k_m is the model quant type, q3_k_s, q4_k_m, ..., can also be used)
C:\Users\developer>ollama run hf.co/John1604/MiniMax-M2.1-gguf:q3_k_m
pulling manifest
...
writing manifest
success
>>> Send a message (/? for help)
- After you run command: ollama run hf.co/John1604/MiniMax-M2.1-gguf:q3_k_m, it will appear in ollama UI - you may select this model hf.co/John1604/MiniMax-M2.1-gguf:q3_k_m from the model list, and run it the same way as other ollama supported models.
Use the model in LM Studio
- download and install LM Studio https://lmstudio.ai/
Discover models
- In the LM Studio, click "Discover" icon. "Mission Control" popup window will be displayed.
- In the "Mission Control" search bar, type "John1604/MiniMax-M2.1-gguf" and check "GGUF", the model should be found.
- Download a quantized model.
- Load the quantized model.
- Ask questions.
quantized models comparison
| Type | Bits | Quality | Description |
|---|---|---|---|
| Q2_K | 2-bit | π₯ Low | Minimal footprint; only for tests |
| Q3_K_S | 3-bit | π§ Low | βSmallβ variant (less accurate) |
| Q3_K_M | 3-bit | π§ LowβMed | βMediumβ variant |
| Q4_K_S | 4-bit | π¨ Med | Small, faster, slightly less quality |
| Q4_K_M | 4-bit | π© MedβHigh | βMediumβ β best 4-bit balance |
| Q5_K_S | 5-bit | π© High | Slightly smaller than Q5_K_M |
| Q5_K_M | 5-bit | π©π© High | Excellent general-purpose quant |
| Q6_K | 6-bit | π©π©π© Very High | Almost FP16 quality, larger size |
| Q8_0 | 8-bit | π©π©π©π© | Near-lossless baseline |
- Downloads last month
- 193
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for John1604/MiniMax-M2.1-gguf
Base model
MiniMaxAI/MiniMax-M2.1