Qwen3.6-35B-A3B - Q6_K GGUF Quantization

This repository contains the Q6_K GGUF format of the Qwen3.6-35B-A3B model.

These files were quantized by Abiray using llama.cpp to make the model accessible for consumer hardware and CPU-heavy environments.

πŸ“¦ Other Available Formats

I have processed this model into several different quantization formats. You can find them in my other repositories:

πŸ’» How to run with llama.cpp

You can run this model locally using llama-cli from the llama.cpp project.

# Example command (adjust threads and context size to your machine)
./llama-cli -m Qwen3.6-35B-A3B-Q6_K.gguf -p "Your prompt here" -n 512 -t 8 -c 4096
Downloads last month
827
GGUF
Model size
35B params
Architecture
qwen35moe
Hardware compatibility
Log In to add your hardware

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Abiray/Qwen3.6-35B-A3B-Q6_K-GGUF

Quantized
(161)
this model

Collection including Abiray/Qwen3.6-35B-A3B-Q6_K-GGUF