Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ language:
|
|
11 |
- en
|
12 |
---
|
13 |
|
14 |
-
# Uploaded model - AlphaAI-
|
15 |
|
16 |
- **Developed by:** alphaaico
|
17 |
- **License:** apache-2.0
|
@@ -21,7 +21,7 @@ This llama model was trained 2x faster with [Unsloth](https://github.com/unsloth
|
|
21 |
|
22 |
**Overview**
|
23 |
|
24 |
-
AlphaAI-
|
25 |
|
26 |
Designed for local AI deployments, it supports efficient inference on personal hardware while maintaining high reasoning capabilities. The training process was accelerated using Unsloth and Hugging Face's TRL library, allowing for 2x faster fine-tuning.
|
27 |
|
@@ -57,7 +57,7 @@ https://huggingface.co/alphaaico/AAI-1.5B-Thought
|
|
57 |
|
58 |
**Limitations & Biases**
|
59 |
|
60 |
-
As with any AI model, AlphaAI-
|
61 |
|
62 |
**Acknowledgments**
|
63 |
|
|
|
11 |
- en
|
12 |
---
|
13 |
|
14 |
+
# Uploaded model - AlphaAI-1.5B-Thought
|
15 |
|
16 |
- **Developed by:** alphaaico
|
17 |
- **License:** apache-2.0
|
|
|
21 |
|
22 |
**Overview**
|
23 |
|
24 |
+
AlphaAI-1.5B-Thought is a fine-tuned version of Qwen2.5-1.5B, optimized for chain-of-thought (CoT) reasoning and structured problem-solving. This model has been trained on a custom CoT dataset, enhancing its ability to perform step-by-step logical reasoning, multi-step inference, and contextual understanding across various domains.
|
25 |
|
26 |
Designed for local AI deployments, it supports efficient inference on personal hardware while maintaining high reasoning capabilities. The training process was accelerated using Unsloth and Hugging Face's TRL library, allowing for 2x faster fine-tuning.
|
27 |
|
|
|
57 |
|
58 |
**Limitations & Biases**
|
59 |
|
60 |
+
As with any AI model, AlphaAI-1.5B-Thought may reflect biases present in its training data. Users should validate responses for critical applications and fine-tune further for domain-specific tasks.
|
61 |
|
62 |
**Acknowledgments**
|
63 |
|