ape = Algorithmic Pattern Emulation

A finetune of Mistral Nemo Instruct 2407 using my fullfictions-85kmax dataset. The prompts are fairly simple. My goal is to train a model that can write long ficiton that make sense. The training data contains the full text of public domain short stories and novels. 85k of context is about the limit I've been able to train without getting OOM errors using rented GPUs.

Thanks to unsloth brothers, gutenberg volunteers, the Mistral Nemo team, and the folks in various discord servers who have helped me out.

Downloads last month
7
Safetensors
Model size
12.2B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for leftyfeep/ape-fiction-full-instruct

Quantizations
1 model