feizhengcong commited on
Commit
9c456d8
·
verified ·
1 Parent(s): 9346a4a

Upload config.json

Browse files
Files changed (1) hide show
  1. transformer/config.json +36 -0
transformer/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LFE_heads": 12,
3
+ "LFE_num_tokens": 32,
4
+ "LFE_output_dim": 768,
5
+ "_class_name": "ConsisIDTransformer3DModel",
6
+ "_diffusers_version": "0.31.0",
7
+ "activation_fn": "gelu-approximate",
8
+ "attention_bias": true,
9
+ "attention_head_dim": 64,
10
+ "cross_attn_interval": 2,
11
+ "dropout": 0.0,
12
+ "flip_sin_to_cos": true,
13
+ "freq_shift": 0,
14
+ "in_channels": 32,
15
+ "is_kps": false,
16
+ "is_train_face": true,
17
+ "local_face_scale": 1.0,
18
+ "max_text_seq_length": 226,
19
+ "norm_elementwise_affine": true,
20
+ "norm_eps": 1e-05,
21
+ "num_attention_heads": 48,
22
+ "num_layers": 42,
23
+ "out_channels": 16,
24
+ "patch_size": 2,
25
+ "sample_frames": 49,
26
+ "sample_height": 60,
27
+ "sample_width": 90,
28
+ "spatial_interpolation_scale": 1.875,
29
+ "temporal_compression_ratio": 4,
30
+ "temporal_interpolation_scale": 1.0,
31
+ "text_embed_dim": 4096,
32
+ "time_embed_dim": 512,
33
+ "timestep_activation_fn": "silu",
34
+ "use_learned_positional_embeddings": true,
35
+ "use_rotary_positional_embeddings": true
36
+ }