🍷 FineLlama-3.1-8B

This is a finetune of meta-llama/Meta-Llama-3.1-8B made for my article "Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth".

It was trained on 100k super high-quality samples from the mlabonne/FineTome-100k dataset.

Try the demo: https://huggingface.co/spaces/mlabonne/FineLlama-3.1-8B

πŸ”Ž Applications

This model was made for educational purposes. I recommend using Meta's instruct model for real applications.

⚑ Quantization

πŸ† Evaluation

TBD.

πŸ’» Usage

!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "mlabonne/FineLlama-3.1-8B"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
3,133
Safetensors
Model size
8.03B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for mlabonne/FineLlama-3.1-8B

Finetuned
(603)
this model
Quantizations
10 models

Dataset used to train mlabonne/FineLlama-3.1-8B

Spaces using mlabonne/FineLlama-3.1-8B 7