nhatminh commited on
Commit
ea8042d
·
verified ·
1 Parent(s): 5ddaf2a

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +8 -17
config.json CHANGED
@@ -1,13 +1,11 @@
1
  {
2
- "_name_or_path": "/home/s48gb/Desktop/minhndn/Finetune-Jina-rerank/jina_v2",
3
- "architectures": [
4
- "XLMRobertaForSequenceClassification"
5
- ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "auto_map": {
8
  "AutoConfig": "configuration_xlm_roberta.XLMRobertaFlashConfig",
9
- "AutoModel": "jinaai/jina-reranker-v2-base-multilingual--modeling_xlm_roberta.XLMRobertaModel",
10
- "AutoModelForSequenceClassification": "jinaai/jina-reranker-v2-base-multilingual--modeling_xlm_roberta.XLMRobertaForSequenceClassification"
11
  },
12
  "bos_token_id": 0,
13
  "classifier_dropout": null,
@@ -16,6 +14,7 @@
16
  "hidden_act": "gelu",
17
  "hidden_dropout_prob": 0.1,
18
  "hidden_size": 768,
 
19
  "id2label": {
20
  "0": "LABEL_0"
21
  },
@@ -24,14 +23,7 @@
24
  "label2id": {
25
  "LABEL_0": 0
26
  },
27
- "layer_norm_eps": 1e-05,
28
- "load_trained_adapters": false,
29
- "lora_adaptations": null,
30
- "lora_alpha": 1,
31
- "lora_dropout_p": 0.0,
32
- "lora_main_params_trainable": false,
33
- "lora_rank": 4,
34
- "matryoshka_dimensions": null,
35
  "max_position_embeddings": 1026,
36
  "num_attention_heads": 12,
37
  "num_hidden_layers": 12,
@@ -39,10 +31,9 @@
39
  "pad_token_id": 1,
40
  "position_embedding_type": "absolute",
41
  "torch_dtype": "bfloat16",
42
- "transformers_version": "4.45.2",
43
- "truncate_dim": null,
44
  "type_vocab_size": 1,
45
  "use_cache": false,
46
  "use_flash_attn": true,
47
  "vocab_size": 250002
48
- }
 
1
  {
2
+ "_name_or_path": "jinaai/jina-reranker-v2-base-multilingual",
3
+ "architectures": ["XLMRobertaForSequenceClassification"],
 
 
4
  "attention_probs_dropout_prob": 0.1,
5
  "auto_map": {
6
  "AutoConfig": "configuration_xlm_roberta.XLMRobertaFlashConfig",
7
+ "AutoModel": "modeling_xlm_roberta.XLMRobertaModel",
8
+ "AutoModelForSequenceClassification": "modeling_xlm_roberta.XLMRobertaForSequenceClassification"
9
  },
10
  "bos_token_id": 0,
11
  "classifier_dropout": null,
 
14
  "hidden_act": "gelu",
15
  "hidden_dropout_prob": 0.1,
16
  "hidden_size": 768,
17
+ "num_labels": 1,
18
  "id2label": {
19
  "0": "LABEL_0"
20
  },
 
23
  "label2id": {
24
  "LABEL_0": 0
25
  },
26
+ "layer_norm_eps": 1e-5,
 
 
 
 
 
 
 
27
  "max_position_embeddings": 1026,
28
  "num_attention_heads": 12,
29
  "num_hidden_layers": 12,
 
31
  "pad_token_id": 1,
32
  "position_embedding_type": "absolute",
33
  "torch_dtype": "bfloat16",
34
+ "transformers_version": "4.40.0",
 
35
  "type_vocab_size": 1,
36
  "use_cache": false,
37
  "use_flash_attn": true,
38
  "vocab_size": 250002
39
+ }