File size: 486 Bytes
4d34db4 |
1 |
{"seed": 51, "batch_size": 1024, "buffer_mult": 512, "lr": 2e-05, "num_tokens": 2136820, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 768, "d_model": 768, "dict_size": 65536, "seq_len": 1024, "enc_dtype": "fp32", "model_name": "gpt2-small", "site": "resid_post", "device": "cuda:1", "model_batch_size": 32, "log_every": 100, "save_every": 100000, "dec_init_norm": 0.08, "hook_point": "blocks.11.hook_resid_post", "wandb_project": "crosscoder", "wandb_entity": "aaronjerez-byte"} |