alphaaico commited on
Commit
5f9d08c
·
verified ·
1 Parent(s): 0d56ef0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -11,7 +11,7 @@ language:
11
  - en
12
  ---
13
 
14
- # Uploaded model - AlphaAI-CoT-Reasoner-1.5B
15
 
16
  - **Developed by:** alphaaico
17
  - **License:** apache-2.0
@@ -21,7 +21,7 @@ This llama model was trained 2x faster with [Unsloth](https://github.com/unsloth
21
 
22
  **Overview**
23
 
24
- AlphaAI-CoT-Reasoner-1.5B is a fine-tuned version of Qwen2.5-1.5B, optimized for chain-of-thought (CoT) reasoning and structured problem-solving. This model has been trained on a custom CoT dataset, enhancing its ability to perform step-by-step logical reasoning, multi-step inference, and contextual understanding across various domains.
25
 
26
  Designed for local AI deployments, it supports efficient inference on personal hardware while maintaining high reasoning capabilities. The training process was accelerated using Unsloth and Hugging Face's TRL library, allowing for 2x faster fine-tuning.
27
 
@@ -57,7 +57,7 @@ https://huggingface.co/alphaaico/AAI-1.5B-Thought
57
 
58
  **Limitations & Biases**
59
 
60
- As with any AI model, AlphaAI-CoT-Reasoner-1.5B may reflect biases present in its training data. Users should validate responses for critical applications and fine-tune further for domain-specific tasks.
61
 
62
  **Acknowledgments**
63
 
 
11
  - en
12
  ---
13
 
14
+ # Uploaded model - AlphaAI-1.5B-Thought
15
 
16
  - **Developed by:** alphaaico
17
  - **License:** apache-2.0
 
21
 
22
  **Overview**
23
 
24
+ AlphaAI-1.5B-Thought is a fine-tuned version of Qwen2.5-1.5B, optimized for chain-of-thought (CoT) reasoning and structured problem-solving. This model has been trained on a custom CoT dataset, enhancing its ability to perform step-by-step logical reasoning, multi-step inference, and contextual understanding across various domains.
25
 
26
  Designed for local AI deployments, it supports efficient inference on personal hardware while maintaining high reasoning capabilities. The training process was accelerated using Unsloth and Hugging Face's TRL library, allowing for 2x faster fine-tuning.
27
 
 
57
 
58
  **Limitations & Biases**
59
 
60
+ As with any AI model, AlphaAI-1.5B-Thought may reflect biases present in its training data. Users should validate responses for critical applications and fine-tune further for domain-specific tasks.
61
 
62
  **Acknowledgments**
63