Add library name tag
Browse filesThis PR ensures the "how to use" button appears on the top right (with a Transformers code snippet).
README.md
CHANGED
@@ -1,12 +1,13 @@
|
|
1 |
---
|
2 |
-
|
|
|
3 |
language:
|
4 |
- en
|
|
|
5 |
metrics:
|
6 |
- accuracy
|
7 |
-
base_model:
|
8 |
-
- meta-llama/Llama-3.1-405B-Instruct
|
9 |
pipeline_tag: text-generation
|
|
|
10 |
---
|
11 |
|
12 |
# CoALM-405B: The Largest Open-Source Agentic LLM
|
@@ -48,8 +49,7 @@ It is designed to **push the boundaries** of open-source agentic LLMs, excelling
|
|
48 |
- **π¨ Largest Open-Source Agentic LLM:** A **405B** parameter model that brings state-of-the-art agentic capabilities to the public domain.
|
49 |
- **π― Best Open-Source Performance on BFCL V3:** Outperforms leading proprietary models like **GPT-4o, Gemini, and Claude** in function-calling tasks.
|
50 |
- **π True Zero-Shot Function Calling:** Generalizes to unseen API tasks with **unmatched accuracy**.
|
51 |
-
- **π€ Multi-Turn Dialogue Mastery:** Excels at long conversations, **task tracking, and complex reasoning
|
52 |
-
- **π API Tool Use and Reasoning:** Makes precise API calls, interprets responses, and synthesizes **coherent** multi-step solutions.
|
53 |
- **π Fully Open-Source & Reproducible:** Released under **cc-by-nc-4.0**, including model weights, training logs, and datasets.
|
54 |
|
55 |
|
@@ -65,8 +65,7 @@ It is designed to **push the boundaries** of open-source agentic LLMs, excelling
|
|
65 |
---
|
66 |
## π§ Training Process
|
67 |
### Fine-tuning Stages
|
68 |
-
1. **TOD Fine-tuning:** Optimized for **dialogue state tracking** (e.g., augmented SNIPS in instruction-tuned format).
|
69 |
-
2. **Function Calling Fine-tuning:** Trained to generate **highly accurate API calls** from LA datasets.
|
70 |
3. **ReAct-based Fine-tuning:** Enhances multi-turn conversations with structured **thought-action-observation-response reasoning**.
|
71 |
|
72 |
### Training Hyperparameters
|
@@ -128,7 +127,4 @@ If you use **CoALM-405B** in your research, please cite:
|
|
128 |
}
|
129 |
```
|
130 |
|
131 |
-
For more details, visit [Project Repository](https://github.com/oumi-ai/oumi/tree/main/configs/projects/CALM) or contact **[email protected]**.
|
132 |
-
|
133 |
-
|
134 |
-
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- meta-llama/Llama-3.1-405B-Instruct
|
4 |
language:
|
5 |
- en
|
6 |
+
license: cc-by-nc-4.0
|
7 |
metrics:
|
8 |
- accuracy
|
|
|
|
|
9 |
pipeline_tag: text-generation
|
10 |
+
library_name: transformers
|
11 |
---
|
12 |
|
13 |
# CoALM-405B: The Largest Open-Source Agentic LLM
|
|
|
49 |
- **π¨ Largest Open-Source Agentic LLM:** A **405B** parameter model that brings state-of-the-art agentic capabilities to the public domain.
|
50 |
- **π― Best Open-Source Performance on BFCL V3:** Outperforms leading proprietary models like **GPT-4o, Gemini, and Claude** in function-calling tasks.
|
51 |
- **π True Zero-Shot Function Calling:** Generalizes to unseen API tasks with **unmatched accuracy**.
|
52 |
+
- **π€ Multi-Turn Dialogue Mastery:** Excels at long conversations, **task tracking, and complex reasoning**.\n- **π API Tool Use and Reasoning:** Makes precise API calls, interprets responses, and synthesizes **coherent** multi-step solutions.
|
|
|
53 |
- **π Fully Open-Source & Reproducible:** Released under **cc-by-nc-4.0**, including model weights, training logs, and datasets.
|
54 |
|
55 |
|
|
|
65 |
---
|
66 |
## π§ Training Process
|
67 |
### Fine-tuning Stages
|
68 |
+
1. **TOD Fine-tuning:** Optimized for **dialogue state tracking** (e.g., augmented SNIPS in instruction-tuned format).\n2. **Function Calling Fine-tuning:** Trained to generate **highly accurate API calls** from LA datasets.
|
|
|
69 |
3. **ReAct-based Fine-tuning:** Enhances multi-turn conversations with structured **thought-action-observation-response reasoning**.
|
70 |
|
71 |
### Training Hyperparameters
|
|
|
127 |
}
|
128 |
```
|
129 |
|
130 |
+
For more details, visit [Project Repository](https://github.com/oumi-ai/oumi/tree/main/configs/projects/CALM) or contact **[email protected]**.
|
|
|
|
|
|