akhazr commited on
Commit
b609478
·
verified ·
1 Parent(s): 158f3c5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -4
README.md CHANGED
@@ -8,14 +8,22 @@ base_model: meta-llama/Meta-Llama-3-8B
8
 
9
  ## Overview
10
 
11
- **nephra v1.0** is a model built for roleplaying sessions, trained on roleplay and instruction-style datasets.
 
 
 
 
 
 
 
 
12
 
13
 
14
  ```python
15
  import transformers
16
  import torch
17
 
18
- model_id = "yodayo-ai/nephra_v1.0"
19
 
20
  pipeline = transformers.pipeline(
21
  "text-generation",
@@ -51,7 +59,7 @@ print(outputs[0]["generated_text"][len(prompt):])
51
 
52
  ### Recommended Settings
53
 
54
- To guide the model towards generating high quality responses, here are the ideal settings:
55
 
56
  ```
57
  Prompt Format: Same Prompt Format as Llama-3-Instruct
@@ -62,5 +70,30 @@ Custom Stopping Strings: "\n{{user}}", "<" , "```" , -> Has occasional broken ge
62
  ```
63
 
64
 
65
- nephra v1 falls under [META LLAMA 3 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
 
 
 
8
 
9
  ## Overview
10
 
11
+ **nephra v1** is primarily a model built for roleplaying sessions, trained on roleplay and instruction-style datasets.
12
+
13
+ ## Model Details
14
+ - **Developed by**: [Sao10K](https://huggingface.co/Sao10K)
15
+ - **Model type**: Text-based Large Language Model
16
+ - **License**: [Meta Llama 3 Community License Agreement](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE)
17
+ - **Finetuned from model**: [Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
18
+
19
+ ## Inference Guidelines
20
 
21
 
22
  ```python
23
  import transformers
24
  import torch
25
 
26
+ model_id = "yodayo-ai/Nephra_V1.0"
27
 
28
  pipeline = transformers.pipeline(
29
  "text-generation",
 
59
 
60
  ### Recommended Settings
61
 
62
+ To guide the model to generate high-quality responses, here are the ideal settings:
63
 
64
  ```
65
  Prompt Format: Same Prompt Format as Llama-3-Instruct
 
70
  ```
71
 
72
 
73
+ ## Training
74
+
75
+ These are the key hyperparameters used during training:
76
+
77
+ | Hyperparameters | Finetuning |
78
+ |-------------------------------|----------------------------|
79
+ | **Hardware** | 4x Nvidia L40 48GB |
80
+ | **Batch Size** | 4x 2 |
81
+ | **Gradient Accumulation Steps** | 4x 3 |
82
+ | **LoRA Rank** | 32 |
83
+ | **LoRA Alpha** | 64 |
84
+ | **LoRA Dropout** | 0.04 |
85
+ | **Seq_Length** | 8192 |
86
+ | **LoRA Target Layers** | All Linear Layers |
87
+ | **Epochs** | 2 |
88
+ | **Max Learning Rate** | 2e-4 |
89
+ | **Min Learning Rate** | 4e-5 |
90
+ | **Optimizer** | adamw_bnb_8bit |
91
+ | **Optimizer Args** | Warmup: True | Steps: 20
92
+ | **Scheduler** | cosine_with_min_lr |
93
+ | **Warmup Steps** | 4% |
94
+
95
+ ## License
96
+
97
+ Nephra v1 falls under [META LLAMA 3 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta-llama/Meta-Llama-3-8B/blob/main/LICENSE).
98
 
99
+ to