Update README.md
Browse files
README.md
CHANGED
@@ -8,14 +8,22 @@ base_model: meta-llama/Meta-Llama-3-8B
|
|
8 |
|
9 |
## Overview
|
10 |
|
11 |
-
**nephra v1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
|
13 |
|
14 |
```python
|
15 |
import transformers
|
16 |
import torch
|
17 |
|
18 |
-
model_id = "yodayo-ai/
|
19 |
|
20 |
pipeline = transformers.pipeline(
|
21 |
"text-generation",
|
@@ -51,7 +59,7 @@ print(outputs[0]["generated_text"][len(prompt):])
|
|
51 |
|
52 |
### Recommended Settings
|
53 |
|
54 |
-
To guide the model
|
55 |
|
56 |
```
|
57 |
Prompt Format: Same Prompt Format as Llama-3-Instruct
|
@@ -62,5 +70,30 @@ Custom Stopping Strings: "\n{{user}}", "<" , "```" , -> Has occasional broken ge
|
|
62 |
```
|
63 |
|
64 |
|
65 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66 |
|
|
|
|
8 |
|
9 |
## Overview
|
10 |
|
11 |
+
**nephra v1** is primarily a model built for roleplaying sessions, trained on roleplay and instruction-style datasets.
|
12 |
+
|
13 |
+
## Model Details
|
14 |
+
- **Developed by**: [Sao10K](https://huggingface.co/Sao10K)
|
15 |
+
- **Model type**: Text-based Large Language Model
|
16 |
+
- **License**: [Meta Llama 3 Community License Agreement](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE)
|
17 |
+
- **Finetuned from model**: [Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
|
18 |
+
|
19 |
+
## Inference Guidelines
|
20 |
|
21 |
|
22 |
```python
|
23 |
import transformers
|
24 |
import torch
|
25 |
|
26 |
+
model_id = "yodayo-ai/Nephra_V1.0"
|
27 |
|
28 |
pipeline = transformers.pipeline(
|
29 |
"text-generation",
|
|
|
59 |
|
60 |
### Recommended Settings
|
61 |
|
62 |
+
To guide the model to generate high-quality responses, here are the ideal settings:
|
63 |
|
64 |
```
|
65 |
Prompt Format: Same Prompt Format as Llama-3-Instruct
|
|
|
70 |
```
|
71 |
|
72 |
|
73 |
+
## Training
|
74 |
+
|
75 |
+
These are the key hyperparameters used during training:
|
76 |
+
|
77 |
+
| Hyperparameters | Finetuning |
|
78 |
+
|-------------------------------|----------------------------|
|
79 |
+
| **Hardware** | 4x Nvidia L40 48GB |
|
80 |
+
| **Batch Size** | 4x 2 |
|
81 |
+
| **Gradient Accumulation Steps** | 4x 3 |
|
82 |
+
| **LoRA Rank** | 32 |
|
83 |
+
| **LoRA Alpha** | 64 |
|
84 |
+
| **LoRA Dropout** | 0.04 |
|
85 |
+
| **Seq_Length** | 8192 |
|
86 |
+
| **LoRA Target Layers** | All Linear Layers |
|
87 |
+
| **Epochs** | 2 |
|
88 |
+
| **Max Learning Rate** | 2e-4 |
|
89 |
+
| **Min Learning Rate** | 4e-5 |
|
90 |
+
| **Optimizer** | adamw_bnb_8bit |
|
91 |
+
| **Optimizer Args** | Warmup: True | Steps: 20
|
92 |
+
| **Scheduler** | cosine_with_min_lr |
|
93 |
+
| **Warmup Steps** | 4% |
|
94 |
+
|
95 |
+
## License
|
96 |
+
|
97 |
+
Nephra v1 falls under [META LLAMA 3 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE).
|
98 |
|
99 |
+
to
|