Commit
·
63ef4cb
1
Parent(s):
c205852
Update README.md
Browse files
README.md
CHANGED
@@ -11,16 +11,38 @@ pipeline_tag: text-classification
|
|
11 |
tags:
|
12 |
- textual-entailment
|
13 |
widget:
|
14 |
-
- text: "Qual a capital do Brasil?<s>A capital do Brasil é Brasília!</s>"
|
15 |
example_title: Exemplo
|
16 |
-
- text: "Qual a capital do Brasil?<s>Anões são muito mais legais do que elfos!</s>"
|
17 |
example_title: Exemplo
|
18 |
---
|
19 |
-
# TeenyTinyLlama-162m-
|
20 |
|
21 |
-
TeenyTinyLlama is a series of small foundational models trained in Portuguese.
|
22 |
|
23 |
-
This repository contains a version of [TeenyTinyLlama-162m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-162m) fine-tuned on the [FaQuAD-NLI dataset](https://huggingface.co/datasets/ruanchaves/faquad-nli).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
|
25 |
## Reproducing
|
26 |
|
@@ -115,7 +137,7 @@ trainer.train()
|
|
115 |
|
116 |
```
|
117 |
|
118 |
-
##
|
119 |
|
120 |
| Models | [FaQuAD-NLI](https://huggingface.co/datasets/ruanchaves/faquad-nli) |
|
121 |
|--------------------------------------------------------------------------------------------|---------------------------------------------------------------------|
|
@@ -123,3 +145,26 @@ trainer.train()
|
|
123 |
| [Bert-base-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased) | 93.07 |
|
124 |
| [Gpt2-small-portuguese](https://huggingface.co/pierreguillou/gpt2-small-portuguese) | 86.46 |
|
125 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
tags:
|
12 |
- textual-entailment
|
13 |
widget:
|
14 |
+
- text: "<s>Qual a capital do Brasil?<s>A capital do Brasil é Brasília!</s>"
|
15 |
example_title: Exemplo
|
16 |
+
- text: "<s>Qual a capital do Brasil?<s>Anões são muito mais legais do que elfos!</s>"
|
17 |
example_title: Exemplo
|
18 |
---
|
19 |
+
# TeenyTinyLlama-162m-FaQuAD-NLI
|
20 |
|
21 |
+
TeenyTinyLlama is a series of small foundational models trained in Brazilian Portuguese.
|
22 |
|
23 |
+
This repository contains a version of [TeenyTinyLlama-162m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-162m) (`TeenyTinyLlama-162m-HateBR`) fine-tuned on the [FaQuAD-NLI dataset](https://huggingface.co/datasets/ruanchaves/faquad-nli).
|
24 |
+
|
25 |
+
## Details
|
26 |
+
|
27 |
+
- **Number of Epochs:** 3
|
28 |
+
- **Batch size:** 16
|
29 |
+
- **Optimizer:** `torch.optim.AdamW` (learning_rate = 4e-5, epsilon = 1e-8)
|
30 |
+
- **GPU:** 1 NVIDIA A100-SXM4-40GB
|
31 |
+
|
32 |
+
## Usage
|
33 |
+
|
34 |
+
Using `transformers.pipeline`:
|
35 |
+
|
36 |
+
```python
|
37 |
+
from transformers import pipeline
|
38 |
+
|
39 |
+
text = "<s>Qual a capital do Brasil?<s>A capital do Brasil é Brasília!</s>"
|
40 |
+
|
41 |
+
classifier = pipeline("text-classification", model="nicholasKluge/TeenyTinyLlama-162m-FaQuAD-NLI")
|
42 |
+
classifier(text)
|
43 |
+
|
44 |
+
# >>> [{'label': 'SUITABLE', 'score': 0.9774010181427002}]
|
45 |
+
```
|
46 |
|
47 |
## Reproducing
|
48 |
|
|
|
137 |
|
138 |
```
|
139 |
|
140 |
+
## Fine-Tuning Comparisons
|
141 |
|
142 |
| Models | [FaQuAD-NLI](https://huggingface.co/datasets/ruanchaves/faquad-nli) |
|
143 |
|--------------------------------------------------------------------------------------------|---------------------------------------------------------------------|
|
|
|
145 |
| [Bert-base-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased) | 93.07 |
|
146 |
| [Gpt2-small-portuguese](https://huggingface.co/pierreguillou/gpt2-small-portuguese) | 86.46 |
|
147 |
|
148 |
+
## Cite as 🤗
|
149 |
+
|
150 |
+
```latex
|
151 |
+
|
152 |
+
@misc{nicholas22llama,
|
153 |
+
doi = {10.5281/zenodo.6989727},
|
154 |
+
url = {https://huggingface.co/nicholasKluge/TeenyTinyLlama-162m},
|
155 |
+
author = {Nicholas Kluge Corrêa},
|
156 |
+
title = {TeenyTinyLlama},
|
157 |
+
year = {2023},
|
158 |
+
publisher = {HuggingFace},
|
159 |
+
journal = {HuggingFace repository},
|
160 |
+
}
|
161 |
+
|
162 |
+
```
|
163 |
+
|
164 |
+
## Funding
|
165 |
+
|
166 |
+
This repository was built as part of the RAIES ([Rede de Inteligência Artificial Ética e Segura](https://www.raies.org/)) initiative, a project supported by FAPERGS - ([Fundação de Amparo à Pesquisa do Estado do Rio Grande do Sul](https://fapergs.rs.gov.br/inicial)), Brazil.
|
167 |
+
|
168 |
+
## License
|
169 |
+
|
170 |
+
The TeenyTinyLlama-162m-FaQuAD-NLI is licensed under the Apache License, Version 2.0. See the [LICENSE](LICENSE) file for more details.
|