Caeden-o1 / README.md
caedencode's picture
Update README.md
1edd18f verified
|
raw
history blame
2.5 kB
---
library_name: transformers
tags: [conversational, chain-of-thought, education]
---
# CaedenAI - O1
CaedenAI is a conversational AI model fine-tuned to provide detailed reasoning in its responses using the Chain-of-Thought (CoT) methodology. It is designed for educational use, enabling users to understand the reasoning process behind answers.
## Model Details
### Model Description
- **Developed by:** Caeden Rajoo
- **Model type:** Conversational AI with CoT reasoning
- **License:** Apache 2
- **Finetuned from model:** Qwen/Qwen2.5-1.5B
- **Primary Use Case:** Education and knowledge expansion
This model is fine-tuned for generating step-by-step reasoning for queries, making it an excellent tool for educational environments and learning applications.
## Uses
### Direct Use
This model can be directly applied in:
- Educational environments to help students learn with explanations.
- Applications where detailed reasoning is required for understanding answers.
- Conversational AI systems that prioritize reasoning over simple answers.
### Out-of-Scope Use
This model may not be suitable for:
- Scenarios requiring highly specialized domain knowledge not covered in the training data.
- Tasks requiring real-time response for critical systems (e.g., healthcare, safety).
## Bias, Risks, and Limitations
The model inherits limitations from its training data and base model. Users should consider potential biases or incomplete information in responses.
### Recommendations
- The model's output should be reviewed for accuracy in critical use cases.
- Users should ensure that ethical considerations are met when using the model in sensitive environments.
## How to Get Started with the Model
Here’s how you can load and use CaedenAI:
```python
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("caedencode/Caeden-o1")
tokenizer = AutoTokenizer.from_pretrained("caedencode/Caeden-o1")
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model = model.to(device)
def generate_answer(question):
prompt = f"Question: {question}\nReasoning:\n"
inputs = tokenizer(prompt, return_tensors="pt").to(device)
outputs = model.generate(**inputs, max_length=200, num_beams=5, early_stopping=True)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
question = "What is the largest planet in our solar system?"
answer = generate_answer(question)
print(answer)
```