File size: 3,025 Bytes
43e98e0 61abf5e 9219dc3 61abf5e 9219dc3 61abf5e b79749b b1a1c46 efe9d8e 7cb873b eab135c b1a1c46 eab135c b1a1c46 a50288a b1a1c46 a50288a b1a1c46 a50288a b1a1c46 eab135c b1a1c46 39a6489 b1a1c46 39a6489 b1a1c46 eab135c b1a1c46 a70ad0b b1a1c46 eab135c b1a1c46 b1f0e74 b1a1c46 d718b53 b1a1c46 d718b53 b1a1c46 d718b53 b1a1c46 eab135c b1a1c46 b1f0e74 b1a1c46 b1f0e74 b1a1c46 b1f0e74 b1a1c46 eab135c 7224e57 512947a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 |
---
library_name: transformers
tags:
- math
- lora
- science
- chemistry
- biology
- code
- text-generation-inference
- unsloth
- llama
license: apache-2.0
datasets:
- HuggingFaceTB/smoltalk
language:
- en
- de
- es
- fr
- it
- pt
- hi
- th
base_model:
- meta-llama/Llama-3.2-1B-Instruct
---
![FastLlama-Logo](FastLlama.png)
You can use ChatML & Alpaca format.
You can chat with the model via this [space](https://huggingface.co/spaces/suayptalha/Chat-with-FastLlama).
**Overview:**
FastLlama is a highly optimized version of the Llama-3.2-1B-Instruct model. Designed for superior performance in constrained environments, it combines speed, compactness, and high accuracy. This version has been fine-tuned using the MetaMathQA-50k section of the HuggingFaceTB/smoltalk dataset to enhance its mathematical reasoning and problem-solving abilities.
**Features:**
Lightweight and Fast: Optimized to deliver Llama-class capabilities with reduced computational overhead.
Fine-Tuned for Math Reasoning: Utilizes MetaMathQA-50k for better handling of complex mathematical problems and logical reasoning tasks.
Instruction-Tuned: Pre-trained on instruction-following tasks, making it robust in understanding and executing detailed queries.
Versatile Use Cases: Suitable for educational tools, tutoring systems, or any application requiring mathematical reasoning.
**Performance Highlights:**
Smaller Footprint: The model delivers comparable results to larger counterparts while operating efficiently on smaller hardware.
Enhanced Accuracy: Demonstrates improved performance on mathematical QA benchmarks.
Instruction Adherence: Retains high fidelity in understanding and following user instructions, even for complex queries.
**Loading the Model:**
```py
import torch
from transformers import pipeline
model_id = "suayptalha/FastLlama-3.2-1B-Instruct"
pipe = pipeline(
"text-generation",
model=model_id,
device_map="auto",
)
messages = [
{"role": "system", "content": "You are a friendly assistant named FastLlama."},
{"role": "user", "content": "Who are you?"},
]
outputs = pipe(
messages,
max_new_tokens=256,
)
print(outputs[0]["generated_text"][-1])
```
**Dataset:**
Dataset: MetaMathQA-50k
The MetaMathQA-50k subset of HuggingFaceTB/smoltalk was selected for fine-tuning due to its focus on mathematical reasoning, multi-step problem-solving, and logical inference. The dataset includes:
Algebraic problems
Geometric reasoning tasks
Statistical and probabilistic questions
Logical deduction problems
**Model Fine-Tuning:**
Fine-tuning was conducted using the following configuration:
Learning Rate: 2e-4
Epochs: 1
Optimizer: AdamW
Framework: Unsloth
**License:**
This model is licensed under the Apache 2.0 License. See the LICENSE file for details.
<a href="https://www.buymeacoffee.com/suayptalha" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a> |