LFM2-12B-A1B-SpeedDemon-The-Deckard-II-HERETIC-Uncensored

Fine tune of EXPANDED "LFM2-8B-A1B" to "LFM2-12B-A1B" (almost 50%, custom scripting) using Unsloth using custom dataset(s), 128k context in 16 bit precision.

Expands base model from 24 to 32 layers, 256 to 342 tensors to give the model both more knowledge and brainpower.

Trained on Deckard II [5 datasets, in house]:

Excels at long form creative generation. Very intelligent too.

This is full uncensored version (uncensored first, trained second)

This model is a sparse mixture of experts model (32) with 4 experts activated.

Speed exceeds 50-100 t/s on CPU // 200 t/s on most cards // 400 t/s + on 5090 at QUANT Q6K [4 experts].

One example generation below.

Can also be used on phones // mobile devices.

IN HOUSE BENCHMARKS [by Nightmedia]:

         arc-c arc/e boolq hswag obkqa piqa  wino

LFM2-12B-A1B-The-Deckard-II-HERETIC-Uncensored-Series-A
bf16   	 0.469,0.612,0.782,0.676,0.412,0.743,0.610

---

BASE UNTUNED MODEL:

LFM2-8B-A1B
bf16     0.464,0.583,0.826,0.624,0.398,0.717,0.575
mxfp8    0.460,0.575,0.829,0.624,0.394,0.711,0.567

EXAMPLE GENERATION: [4 experts, Q6K]

Downloads last month
2
Safetensors
Model size
11B params
Tensor type
BF16
·
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DavidAU/LFM2-12B-A1B-SpeedDemon-The-Deckard-II-HERETIC-Uncensored

Finetuned
(19)
this model

Collections including DavidAU/LFM2-12B-A1B-SpeedDemon-The-Deckard-II-HERETIC-Uncensored