laurabernardy commited on
Commit
a9ca440
·
1 Parent(s): 4476a8c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -1
README.md CHANGED
@@ -30,4 +30,24 @@ model-index:
30
  metrics:
31
  - type: "perplexity" # Required. Example: wer. Use metric id from https://hf.co/metrics
32
  value: "45.08" # Required. Example: 20.90
33
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  metrics:
31
  - type: "perplexity" # Required. Example: wer. Use metric id from https://hf.co/metrics
32
  value: "45.08" # Required. Example: 20.90
33
+ ---
34
+
35
+ ---
36
+ ## LuxGPT-2 based GER
37
+ GPT-2 model for Text Generation in luxembourgish language, trained on 711 MB of text data, consisting of RTL.lu news articles, comments, parlament speeches, the luxembourgish Wikipedia, Newscrawl, Webcrawl and subtitles. Created via transfer learning with an English base model, feature space mapping from LB on Base feature space and gradual layer freezing.
38
+ The training took place on a 32 GB Nvidia Tesla V100
39
+ - with One Cycle policy for the learning rate
40
+ - with the help of fastai's LR finder
41
+ - for 53.4 hours
42
+ - for 20 epochs and 7 cycles
43
+ - using the fastai library
44
+
45
+
46
+ ## Usage
47
+ ```python
48
+ from transformers import AutoTokenizer, AutoModelForCausalLM
49
+ tokenizer = AutoTokenizer.from_pretrained("laurabernardy/LuxGPT2basedEN")
50
+ model = AutoModelForCausalLM.from_pretrained("laurabernardy/LuxGPT2basedEN")
51
+ ```
52
+ ## Limitations and Biases
53
+ See the [GPT2 model card](https://huggingface.co/gpt2) for considerations on limitations and bias. See the [GPT2 documentation](https://huggingface.co/transformers/model_doc/gpt2.html) for details on GPT2.