final_model - Merged Model
This model is a merged version of the adapter mekala-2402/model_long with base model unsloth/Llama-3.2-3B-Instruct.
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model = AutoModelForCausalLM.from_pretrained(
"mekala-2402/final_model",
torch_dtype=torch.float16,
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("mekala-2402/final_model")
# Generate text
inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_length=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Model Details
- Original adapter: mekala-2402/model_long
- Base model: unsloth/Llama-3.2-3B-Instruct
- Merged using: PEFT library
GGUF Conversion
To convert this merged model to GGUF format, you can use:
License
This model inherits the license from the base model: unsloth/Llama-3.2-3B-Instruct
- Downloads last month
- 5