|
|
|
--- |
|
language: |
|
- "lb" |
|
license: "mit" |
|
tags: |
|
- "luxembourgish" |
|
- "lëtzebuergesch" |
|
- "text generation" |
|
model-index: |
|
- name: "LuxGPT2" |
|
results: |
|
- task: |
|
type: "text-generation" |
|
name: "Text Generation" |
|
dataset: |
|
type: "LuxembourgishTestDataset" |
|
name: "Luxembourgish Test Dataset" |
|
metrics: |
|
- type: "accuracy" |
|
value: "0.33" |
|
- name: "LuxGPT2" |
|
results: |
|
- task: |
|
type: "text-generation" |
|
name: "Text Generation" |
|
dataset: |
|
type: "LuxembourgishTestDataset" |
|
name: "Luxembourgish Test Dataset" |
|
metrics: |
|
- type: "perplexity" |
|
value: "46.69" |
|
--- |
|
|
|
GPT-2 model for Text Generation in luxembourgish language, trained on 636.8 MB of text data, consisting of RTL.lu news articles, comments, parlament speeches, the luxembourgish Wikipedia, Newscrawl, Webcrawl and subtitles. |
|
The training took place on a 32 GB Nvidia Tesla V100 |
|
with an initial learning rate of 5e-5 |
|
with Batch size 4 |
|
for 109 hours |
|
for 30 epochs |
|
|
|
|
|
See the GPT2 model card for considerations on limitations and bias. See the GPT2 documentation for details on GPT2. |