Jr23xd23's picture
Update README.md
6fa5ed4 verified
---
language:
- ar
tags:
- llama
- text-generation
- instruct
- arabic
- math
- fine-tuned
datasets:
- Jr23xd23/Arabic_LLaMA_Math_Dataset
license: apache-2.0
base_model: meta-llama/Llama-3.2-3B-Instruct
pipeline_tag: text-generation
inference: true
---
# Math_Arabic_Llama-3.2-3B-Instruct
## Model Description
**Math_Arabic_Llama-3.2-3B-Instruct** is a fine-tuned version of the **Llama-3.2-3B-Instruct** model, tailored for solving mathematical problems in Arabic. The model was trained using the **[Arabic LLaMA Math Dataset](https://github.com/jaberjaber23/Arabic-LLaMA-Math-Dataset)**, which includes a wide range of mathematical problems in natural language (Arabic). This model is ideal for educational applications, tutoring, and systems that require automatic math problem-solving in Arabic.
## Model Details
- **Model Type**: Transformer-based language model fine-tuned for text generation
- **Languages**: Arabic
- **Base Model**: [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
- **Dataset**: [Arabic LLaMA Math Dataset](https://github.com/jaberjaber23/Arabic-LLaMA-Math-Dataset)
- **Number of Parameters**: 3 billion
- **Fine-tuned by**: [Jr23xd23](https://huggingface.co/Jr23xd23)
## Training Data
The model was fine-tuned on the **Arabic LLaMA Math Dataset**, which consists of 12,496 examples covering various mathematical topics, such as:
- Basic Arithmetic
- Algebra
- Geometry
- Probability
- Combinatorics
Each example in the dataset includes:
- **Instruction**: The problem statement in Arabic
- **Solution**: The solution to the problem in Arabic
## Intended Use
### Primary Use Cases:
- Solving mathematical problems in Arabic
- Educational applications
- Tutoring systems for Arabic-speaking students
- Mathematical reasoning tasks in Arabic
### How to Use
You can use the model in Python with the Hugging Face transformers library:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct")
model = AutoModelForCausalLM.from_pretrained("Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct")
# Example: Solving a math problem in Arabic
input_text = "ู…ุง ู‡ูˆ ู…ุฌู…ูˆุน ุงู„ุฒูˆุงูŠุง ููŠ ู…ุซู„ุซุŸ" # What is the sum of angles in a triangle?
inputs = tokenizer(input_text, return_tensors="pt")
output = model.generate(**inputs, max_length=100)
print(tokenizer.decode(output[0], skip_special_tokens=True))
```
## Limitations
- The model is not designed for non-mathematical language tasks.
- Performance may degrade when applied to highly complex mathematical problems beyond the scope of the training dataset.
- The model's outputs should be verified for critical applications.
## License
This model is licensed under the **Apache 2.0 License**.
## Citation
If you use this model in your research or projects, please cite it as follows:
```bibtex
@model{Math_Arabic_Llama_3.2_3B_Instruct,
title={Math_Arabic_Llama-3.2-3B-Instruct},
author={Jr23xd23},
year={2024},
publisher={Hugging Face},
url={https://huggingface.co/Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct},
}
```
## Acknowledgements
Special thanks to the creators of the **Arabic LLaMA Math Dataset** for providing a rich resource for fine-tuning the model.