turkeyju commited on
Commit
2d7081c
·
verified ·
1 Parent(s): 49f217d

Push model using huggingface_hub.

Browse files
Files changed (2) hide show
  1. README.md +3 -3
  2. config.json +42 -0
README.md CHANGED
@@ -6,7 +6,7 @@ tags:
6
  - pytorch_model_hub_mixin
7
  - text-to-image-generation
8
  ---
 
9
  This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
10
- - Project page: https://tacju.github.io/projects/maskgen.html
11
- - arXiv: https://arxiv.org/abs/2501.07730
12
- - Library: https://github.com/bytedance/1d-tokenizer
 
6
  - pytorch_model_hub_mixin
7
  - text-to-image-generation
8
  ---
9
+
10
  This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
11
+ - Library: https://github.com/bytedance/1d-tokenizer
12
+ - Docs: [More Information Needed]
 
config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment": {
3
+ "tokenizer_checkpoint": "tatitok_bl32_vae.bin",
4
+ "generator_checkpoint": "maskgen_kl_xl.bin"
5
+ },
6
+ "model": {
7
+ "vq_model": {
8
+ "quantize_mode": "vae",
9
+ "token_size": 16,
10
+ "vit_enc_model_size": "base",
11
+ "vit_dec_model_size": "large",
12
+ "vit_enc_patch_size": 16,
13
+ "vit_dec_patch_size": 16,
14
+ "num_latent_tokens": 32,
15
+ "scale_factor": 0.7525,
16
+ "finetune_decoder": false,
17
+ "is_legacy": false
18
+ },
19
+ "maskgen": {
20
+ "decoder_embed_dim": 1280,
21
+ "decoder_depth": 20,
22
+ "decoder_num_heads": 16,
23
+ "micro_condition": true,
24
+ "micro_condition_embed_dim": 256,
25
+ "text_drop_prob": 0.1,
26
+ "cfg": 3.0,
27
+ "cfg_schedule": "linear",
28
+ "num_iter": 32,
29
+ "temperature": 1.0,
30
+ "sample_aesthetic_score": 6.5
31
+ }
32
+ },
33
+ "losses": {
34
+ "diffloss_d": 8,
35
+ "diffloss_w": 1280
36
+ },
37
+ "dataset": {
38
+ "preprocessing": {
39
+ "crop_size": 256
40
+ }
41
+ }
42
+ }