Ubuntu commited on
Commit
1a552e8
·
1 Parent(s): 67aefb8

Add testing models and malicious content for security research

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,19 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
1
  ---
2
  license: mit
3
+ library_name: transformers
4
+ pipeline_tag: image-text-to-text
5
+ extra_gated_heading: Access Gemma on Hugging Face
6
+ extra_gated_prompt: >-
7
+ To access Gemma on Hugging Face, you’re required to review and agree to
8
+ Google’s usage license. To do this, please ensure you’re logged in to Hugging
9
+ Face and click below. Requests are processed immediately.
10
+ extra_gated_button_content: Acknowledge license
11
+ base_model: Qwen/Qwen2.5-Math-7B
12
+ tags:
13
+ - automatic-speech-recognition
14
+ - automatic-speech-translation
15
+ - audio-text-to-text
16
+ - video-text-to-text
17
  ---
18
+
19
+ ## This is sample README.md file
config.json ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "GptOssForCausalLM"
4
+ ],
5
+ "attention_bias": true,
6
+ "attention_dropout": 0.0,
7
+ "eos_token_id": 200002,
8
+ "experts_per_token": 4,
9
+ "head_dim": 64,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2880,
12
+ "initial_context_length": 4096,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 2880,
15
+ "layer_types": [
16
+ "sliding_attention",
17
+ "full_attention",
18
+ "sliding_attention",
19
+ "full_attention",
20
+ "sliding_attention",
21
+ "full_attention",
22
+ "sliding_attention",
23
+ "full_attention",
24
+ "sliding_attention",
25
+ "full_attention",
26
+ "sliding_attention",
27
+ "full_attention",
28
+ "sliding_attention",
29
+ "full_attention",
30
+ "sliding_attention",
31
+ "full_attention",
32
+ "sliding_attention",
33
+ "full_attention",
34
+ "sliding_attention",
35
+ "full_attention",
36
+ "sliding_attention",
37
+ "full_attention",
38
+ "sliding_attention",
39
+ "full_attention",
40
+ "sliding_attention",
41
+ "full_attention",
42
+ "sliding_attention",
43
+ "full_attention",
44
+ "sliding_attention",
45
+ "full_attention",
46
+ "sliding_attention",
47
+ "full_attention",
48
+ "sliding_attention",
49
+ "full_attention",
50
+ "sliding_attention",
51
+ "full_attention"
52
+ ],
53
+ "max_position_embeddings": 131072,
54
+ "model_type": "gpt_oss",
55
+ "num_attention_heads": 64,
56
+ "num_experts_per_tok": 4,
57
+ "num_hidden_layers": 36,
58
+ "num_key_value_heads": 8,
59
+ "num_local_experts": 128,
60
+ "output_router_logits": false,
61
+ "pad_token_id": 199999,
62
+ "quantization_config": {
63
+ "modules_to_not_convert": [
64
+ "model.layers.*.self_attn",
65
+ "model.layers.*.mlp.router",
66
+ "model.embed_tokens",
67
+ "lm_head"
68
+ ],
69
+ "quant_method": "mxfp4"
70
+ },
71
+ "rms_norm_eps": 1e-05,
72
+ "rope_scaling": {
73
+ "beta_fast": 32.0,
74
+ "beta_slow": 1.0,
75
+ "factor": 32.0,
76
+ "original_max_position_embeddings": 4096,
77
+ "rope_type": "yarn",
78
+ "truncate": false
79
+ },
80
+ "rope_theta": 150000,
81
+ "router_aux_loss_coef": 0.9,
82
+ "sliding_window": 128,
83
+ "swiglu_limit": 7.0,
84
+ "tie_word_embeddings": false,
85
+ "transformers_version": "4.55.0.dev0",
86
+ "use_cache": true,
87
+ "vocab_size": 201088
88
+ }
89
+
eicar-test.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ X5O!P%@AP[4\PZX54(P^)7CC)7}$EICAR-STANDARD-ANTIVIRUS-TEST-FILE!$H+H*
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:add8acd15b36cd028d61d1bc73b6acbc8fe85e28ab1e4c78e9edbf4933b692fd
3
+ size 54
unsafe_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c974d813e452150ef03ff6c871ea1e0b3c305aca35a78dfaee3617a1254a8c5f
3
+ size 498718713
vulnerable_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36cc144d1b6b91ceeff14b9a2bd94be121a40013174c791761a51523d62678c5
3
+ size 26200
vulnerable_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56309527147c0e4a79ad5064e464872c27b1346462c11e7b07c7ea256280905d
3
+ size 390
vulnerable_model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20b4c71d8131655cc725a5228429c4b3c683e3ec42d4cc31b7958abc98245643
3
+ size 1934