cvejoski commited on
Commit
9ee5338
·
verified ·
1 Parent(s): e9cdb01

Upload FIMMJP

Browse files
Files changed (2) hide show
  1. config.json +30 -35
  2. model.safetensors +2 -2
config.json CHANGED
@@ -7,57 +7,52 @@
7
  "AutoModel": "mjp.FIMMJP"
8
  },
9
  "initial_distribution_decoder": {
10
- "dropout": 0.1,
 
 
 
11
  "hidden_layers": [
12
- 64,
13
- 64
14
  ],
15
- "name": "fim.models.blocks.MLP"
 
 
16
  },
17
  "intensity_matrix_decoder": {
18
- "dropout": 0.1,
 
 
 
19
  "hidden_layers": [
20
- 64,
21
- 64
22
  ],
23
- "name": "fim.models.blocks.MLP"
 
 
24
  },
25
  "model_type": "fimmjp",
26
  "n_states": 6,
27
  "path_attention": {
28
- "batch_first": true,
29
- "embed_dim": 64,
30
- "name": "torch.nn.MultiheadAttention",
31
- "num_heads": 8
 
32
  },
33
  "pos_encodings": {
34
- "name": "fim.models.blocks.SineTimeEncoding",
35
- "out_features": 64
36
  },
37
  "torch_dtype": "float32",
38
  "transformers_version": "4.46.0",
39
  "ts_encoder": {
40
- "embed_dim": 64,
41
- "name": "fim.models.blocks.base.TransformerEncoder",
42
- "num_layers": 4,
43
- "transformer_block": {
44
- "activation": {
45
- "name": "torch.nn.ReLU"
46
- },
47
- "attention_head": {
48
- "batch_first": true,
49
- "embed_dim": 64,
50
- "name": "torch.nn.MultiheadAttention",
51
- "num_heads": 8
52
- },
53
- "dropout": 0.1,
54
- "ff_dim": 256,
55
- "in_features": 64,
56
- "name": "fim.models.blocks.TransformerBlock",
57
- "normalization": {
58
- "name": "torch.nn.LayerNorm",
59
- "normalized_shape": 64
60
- }
61
  }
62
  },
63
  "use_adjacency_matrix": false
 
7
  "AutoModel": "mjp.FIMMJP"
8
  },
9
  "initial_distribution_decoder": {
10
+ "dropout": 0,
11
+ "hidden_act": {
12
+ "name": "torch.nn.SELU"
13
+ },
14
  "hidden_layers": [
15
+ 128,
16
+ 128
17
  ],
18
+ "in_features": 2049,
19
+ "initialization_scheme": "lecun_normal",
20
+ "name": "fim.models.blocks.base.MLP"
21
  },
22
  "intensity_matrix_decoder": {
23
+ "dropout": 0,
24
+ "hidden_act": {
25
+ "name": "torch.nn.SELU"
26
+ },
27
  "hidden_layers": [
28
+ 128,
29
+ 128
30
  ],
31
+ "in_features": 2049,
32
+ "initialization_scheme": "lecun_normal",
33
+ "name": "fim.models.blocks.base.MLP"
34
  },
35
  "model_type": "fimmjp",
36
  "n_states": 6,
37
  "path_attention": {
38
+ "embed_dim": 512,
39
+ "kv_dim": 128,
40
+ "n_heads": 1,
41
+ "n_queries": 16,
42
+ "name": "fim.models.blocks.MultiHeadLearnableQueryAttention"
43
  },
44
  "pos_encodings": {
45
+ "name": "fim.models.blocks.positional_encodings.DeltaTimeEncoding"
 
46
  },
47
  "torch_dtype": "float32",
48
  "transformers_version": "4.46.0",
49
  "ts_encoder": {
50
+ "name": "fim.models.blocks.base.RNNEncoder",
51
+ "rnn": {
52
+ "batch_first": true,
53
+ "bidirectional": true,
54
+ "hidden_size": 256,
55
+ "name": "torch.nn.LSTM"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
  }
57
  },
58
  "use_adjacency_matrix": false
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97b87e5a321a9c57addcee2d03d9cd7f996f67bdc91f67023eaf495b2d3a885a
3
- size 1025616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:932ce52b1ca140a94b92bfe2ec7dea9fa2625c857e146970aa4689a4f361892f
3
+ size 4979384