final_model - Merged Model

This model is a merged version of the adapter mekala-2402/model_long with base model unsloth/Llama-3.2-3B-Instruct.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model = AutoModelForCausalLM.from_pretrained(
    "mekala-2402/final_model",
    torch_dtype=torch.float16,
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("mekala-2402/final_model")

# Generate text
inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_length=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Model Details

  • Original adapter: mekala-2402/model_long
  • Base model: unsloth/Llama-3.2-3B-Instruct
  • Merged using: PEFT library

GGUF Conversion

To convert this merged model to GGUF format, you can use:

  1. llama.cpp convert script
  2. Hugging Face GGUF Space

License

This model inherits the license from the base model: unsloth/Llama-3.2-3B-Instruct

Downloads last month
5
Safetensors
Model size
3.21B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for mekala-2402/final_model

Finetuned
(366)
this model
Quantizations
1 model