Llama 3.1 Merged Adapters

Model Description

This is a merged model combining multiple fine-tuned LoRA adapters using TIES (Task Inference with Expert Selection) merging technique. The model combines the strengths of multiple specialized adapters into a single model.

Base Model

Merged Adapters

The following adapters were merged:

  • llama337 - Specialized for creative writing
  • llama338 - Specialized for logical reasoning
  • llama340 - Specialized for code generation
  • llama344 - Specialized for instruction following
  • llama345 - Specialized for factual knowledge
  • llama346 - Specialized for conversational abilities
  • llama349 - Specialized for problem solving
  • llama350 - Specialized for structured output

Merging Parameters

  • Merging Method: TIES (Task Inference with Expert Selection)
  • Density: 0.2 (controls parameter sparsity)
  • Weights: Equal weighting (1.0 for each adapter)
  • Merge Date: 2025-03-09

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the model and tokenizer
model = AutoModelForCausalLM.from_pretrained("kevin009/llama3-merged-adapters")
tokenizer = AutoTokenizer.from_pretrained("kevin009/llama3-merged-adapters")

# Example usage
prompt = "Write a short story about a robot learning to paint."
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Model Performance

This merged model combines the capabilities of multiple specialized adapters, resulting in improved performance across a variety of tasks compared to individual adapters.

Limitations

  • The model inherits limitations from the base Llama 3.1 model
  • May produce inconsistent outputs for certain edge cases
  • As with all language models, can produce incorrect or misleading information

License

This model is subject to the license of the original Llama 3.1 model.

Downloads last month
8
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.