"FP16" is M1/M2 Apple Silicon only optimization that leads to a very noticeable prompt processing boost. For details and benchmarks see jundot/omlx/issues/604.

Please use deepsweet/Qwen3.6-35B-A3B-MLX-oQ8 if you have M3+ Apple Silicon.

This model was converted to MLX format from Qwen/Qwen3.6-35B-A3B using [oMLX v0.36.0].

Settings:

  • Level: oQ8
  • Sensitivity model: Qwen3.6-35B-A3B-MLX-MXFP4-FP16
  • Text Only: yes
  • Non-quant weight dtype: float16
Downloads last month
1,610
Safetensors
Model size
10B params
Tensor type
F16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for deepsweet/Qwen3.6-35B-A3B-MLX-oQ8-FP16

Quantized
(162)
this model

Collection including deepsweet/Qwen3.6-35B-A3B-MLX-oQ8-FP16