nielsr HF staff commited on
Commit
c340f4a
Β·
verified Β·
1 Parent(s): 073a7ee

Add library name tag

Browse files

This PR ensures the "how to use" button appears on the top right (with a Transformers code snippet).

Files changed (1) hide show
  1. README.md +7 -11
README.md CHANGED
@@ -1,12 +1,13 @@
1
  ---
2
- license: cc-by-nc-4.0
 
3
  language:
4
  - en
 
5
  metrics:
6
  - accuracy
7
- base_model:
8
- - meta-llama/Llama-3.1-405B-Instruct
9
  pipeline_tag: text-generation
 
10
  ---
11
 
12
  # CoALM-405B: The Largest Open-Source Agentic LLM
@@ -48,8 +49,7 @@ It is designed to **push the boundaries** of open-source agentic LLMs, excelling
48
  - **🚨 Largest Open-Source Agentic LLM:** A **405B** parameter model that brings state-of-the-art agentic capabilities to the public domain.
49
  - **🎯 Best Open-Source Performance on BFCL V3:** Outperforms leading proprietary models like **GPT-4o, Gemini, and Claude** in function-calling tasks.
50
  - **πŸ” True Zero-Shot Function Calling:** Generalizes to unseen API tasks with **unmatched accuracy**.
51
- - **πŸ€– Multi-Turn Dialogue Mastery:** Excels at long conversations, **task tracking, and complex reasoning**.
52
- - **πŸ›  API Tool Use and Reasoning:** Makes precise API calls, interprets responses, and synthesizes **coherent** multi-step solutions.
53
  - **πŸ“œ Fully Open-Source & Reproducible:** Released under **cc-by-nc-4.0**, including model weights, training logs, and datasets.
54
 
55
 
@@ -65,8 +65,7 @@ It is designed to **push the boundaries** of open-source agentic LLMs, excelling
65
  ---
66
  ## πŸ”§ Training Process
67
  ### Fine-tuning Stages
68
- 1. **TOD Fine-tuning:** Optimized for **dialogue state tracking** (e.g., augmented SNIPS in instruction-tuned format).
69
- 2. **Function Calling Fine-tuning:** Trained to generate **highly accurate API calls** from LA datasets.
70
  3. **ReAct-based Fine-tuning:** Enhances multi-turn conversations with structured **thought-action-observation-response reasoning**.
71
 
72
  ### Training Hyperparameters
@@ -128,7 +127,4 @@ If you use **CoALM-405B** in your research, please cite:
128
  }
129
  ```
130
 
131
- For more details, visit [Project Repository](https://github.com/oumi-ai/oumi/tree/main/configs/projects/CALM) or contact **[email protected]**.
132
-
133
-
134
-
 
1
  ---
2
+ base_model:
3
+ - meta-llama/Llama-3.1-405B-Instruct
4
  language:
5
  - en
6
+ license: cc-by-nc-4.0
7
  metrics:
8
  - accuracy
 
 
9
  pipeline_tag: text-generation
10
+ library_name: transformers
11
  ---
12
 
13
  # CoALM-405B: The Largest Open-Source Agentic LLM
 
49
  - **🚨 Largest Open-Source Agentic LLM:** A **405B** parameter model that brings state-of-the-art agentic capabilities to the public domain.
50
  - **🎯 Best Open-Source Performance on BFCL V3:** Outperforms leading proprietary models like **GPT-4o, Gemini, and Claude** in function-calling tasks.
51
  - **πŸ” True Zero-Shot Function Calling:** Generalizes to unseen API tasks with **unmatched accuracy**.
52
+ - **πŸ€– Multi-Turn Dialogue Mastery:** Excels at long conversations, **task tracking, and complex reasoning**.\n- **πŸ›  API Tool Use and Reasoning:** Makes precise API calls, interprets responses, and synthesizes **coherent** multi-step solutions.
 
53
  - **πŸ“œ Fully Open-Source & Reproducible:** Released under **cc-by-nc-4.0**, including model weights, training logs, and datasets.
54
 
55
 
 
65
  ---
66
  ## πŸ”§ Training Process
67
  ### Fine-tuning Stages
68
+ 1. **TOD Fine-tuning:** Optimized for **dialogue state tracking** (e.g., augmented SNIPS in instruction-tuned format).\n2. **Function Calling Fine-tuning:** Trained to generate **highly accurate API calls** from LA datasets.
 
69
  3. **ReAct-based Fine-tuning:** Enhances multi-turn conversations with structured **thought-action-observation-response reasoning**.
70
 
71
  ### Training Hyperparameters
 
127
  }
128
  ```
129
 
130
+ For more details, visit [Project Repository](https://github.com/oumi-ai/oumi/tree/main/configs/projects/CALM) or contact **[email protected]**.