Ebtihal commited on
Commit
481d3f4
·
1 Parent(s): 6e021b4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -0
README.md CHANGED
@@ -6,6 +6,7 @@ widget:
6
  - text: " السلام عليكم ورحمة[MASK] وبركاتة"
7
  - example_title: "Example 1"
8
  ---
 
9
 
10
  **AraBERTMo** is an Arabic pre-trained language model based on [Google's BERT architechture](https://github.com/google-research/bert). AraBERTMo uses the same BERT-Base config.
11
 
@@ -13,6 +14,14 @@ AraBERTMo now comes in 10 new variants
13
  All models are available on the `HuggingFace` model page under the [Ebtihal](https://huggingface.co/Ebtihal/) name. Checkpoints are available in PyTorch formats.
14
 
15
 
 
 
 
 
 
 
 
 
16
  ## Load Pretrained Model
17
 
18
  You can use this model by installing `torch` or `tensorflow` and Huggingface library `transformers`. And you can use it directly by initializing it like this:
 
6
  - text: " السلام عليكم ورحمة[MASK] وبركاتة"
7
  - example_title: "Example 1"
8
  ---
9
+ # Arabic BERT Mini Model
10
 
11
  **AraBERTMo** is an Arabic pre-trained language model based on [Google's BERT architechture](https://github.com/google-research/bert). AraBERTMo uses the same BERT-Base config.
12
 
 
14
  All models are available on the `HuggingFace` model page under the [Ebtihal](https://huggingface.co/Ebtihal/) name. Checkpoints are available in PyTorch formats.
15
 
16
 
17
+ ## Pretraining Corpus
18
+
19
+ `AraBertMo_base_V1' model was pretrained on ~3 million words:
20
+
21
+ - Arabic version of [OSCAR](https://traces1.inria.fr/oscar/) -
22
+
23
+
24
+
25
  ## Load Pretrained Model
26
 
27
  You can use this model by installing `torch` or `tensorflow` and Huggingface library `transformers`. And you can use it directly by initializing it like this: