Yousefmd commited on
Commit
fd299c8
·
1 Parent(s): d3b730c

Model save

Browse files
README.md CHANGED
@@ -2,8 +2,6 @@
2
  base_model: aubmindlab/bert-large-arabertv02-twitter
3
  tags:
4
  - generated_from_trainer
5
- metrics:
6
- - accuracy
7
  model-index:
8
  - name: arabert-sentiment-classification
9
  results: []
@@ -15,10 +13,6 @@ should probably proofread and complete it, then remove this comment. -->
15
  # arabert-sentiment-classification
16
 
17
  This model is a fine-tuned version of [aubmindlab/bert-large-arabertv02-twitter](https://huggingface.co/aubmindlab/bert-large-arabertv02-twitter) on an unknown dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.5406
20
- - Macro F1: 0.6704
21
- - Accuracy: 0.7990
22
 
23
  ## Model description
24
 
@@ -47,14 +41,6 @@ The following hyperparameters were used during training:
47
  - lr_scheduler_type: linear
48
  - num_epochs: 2
49
 
50
- ### Training results
51
-
52
- | Training Loss | Epoch | Step | Validation Loss | Macro F1 | Accuracy |
53
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|
54
- | No log | 1.0 | 497 | 0.5438 | 0.6447 | 0.7997 |
55
- | 0.6091 | 2.0 | 994 | 0.5406 | 0.6704 | 0.7990 |
56
-
57
-
58
  ### Framework versions
59
 
60
  - Transformers 4.33.3
 
2
  base_model: aubmindlab/bert-large-arabertv02-twitter
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: arabert-sentiment-classification
7
  results: []
 
13
  # arabert-sentiment-classification
14
 
15
  This model is a fine-tuned version of [aubmindlab/bert-large-arabertv02-twitter](https://huggingface.co/aubmindlab/bert-large-arabertv02-twitter) on an unknown dataset.
 
 
 
 
16
 
17
  ## Model description
18
 
 
41
  - lr_scheduler_type: linear
42
  - num_epochs: 2
43
 
 
 
 
 
 
 
 
 
44
  ### Framework versions
45
 
46
  - Transformers 4.33.3
config.json CHANGED
@@ -10,18 +10,18 @@
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 1024,
12
  "id2label": {
13
- "0": "LABEL_0",
14
- "1": "LABEL_1",
15
- "2": "LABEL_2",
16
- "3": "LABEL_3"
17
  },
18
  "initializer_range": 0.02,
19
  "intermediate_size": 4096,
20
  "label2id": {
21
- "LABEL_0": 0,
22
- "LABEL_1": 1,
23
- "LABEL_2": 2,
24
- "LABEL_3": 3
25
  },
26
  "layer_norm_eps": 1e-12,
27
  "max_position_embeddings": 512,
@@ -30,7 +30,6 @@
30
  "num_hidden_layers": 24,
31
  "pad_token_id": 0,
32
  "position_embedding_type": "absolute",
33
- "problem_type": "single_label_classification",
34
  "torch_dtype": "float32",
35
  "transformers_version": "4.33.3",
36
  "type_vocab_size": 2,
 
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 1024,
12
  "id2label": {
13
+ "0": "Positive",
14
+ "1": "Negative",
15
+ "2": "Neutral",
16
+ "3": "Mixed"
17
  },
18
  "initializer_range": 0.02,
19
  "intermediate_size": 4096,
20
  "label2id": {
21
+ "Mixed": 3,
22
+ "Negative": 1,
23
+ "Neutral": 2,
24
+ "Positive": 0
25
  },
26
  "layer_norm_eps": 1e-12,
27
  "max_position_embeddings": 512,
 
30
  "num_hidden_layers": 24,
31
  "pad_token_id": 0,
32
  "position_embedding_type": "absolute",
 
33
  "torch_dtype": "float32",
34
  "transformers_version": "4.33.3",
35
  "type_vocab_size": 2,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:374d1413b0dda79435bcf2f77930aa3ca8f93a465b46458d6fdd8935a733c466
3
  size 1477844273
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4cab89a3e33b78f1f53626ef0d964f55d9b85a3da1629ec010c241188de48a2
3
  size 1477844273
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_basic_tokenize": true,
5
+ "do_lower_case": false,
6
+ "mask_token": "[MASK]",
7
+ "max_len": 512,
8
+ "model_max_length": 512,
9
+ "never_split": [
10
+ "[بريد]",
11
+ "[مستخدم]",
12
+ "[رابط]"
13
+ ],
14
+ "pad_token": "[PAD]",
15
+ "sep_token": "[SEP]",
16
+ "strip_accents": null,
17
+ "tokenize_chinese_chars": true,
18
+ "tokenizer_class": "BertTokenizer",
19
+ "unk_token": "[UNK]",
20
+ "use_fast": true
21
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff