wlog commited on
Commit
8a023e9
·
verified ·
1 Parent(s): 0387fa5

Upload 11 files

Browse files
cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false}, "batch_size": 4, "grad_acc_steps": 8, "micro_acc_steps": 2, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["gpt_neox.layers.0", "gpt_neox.layers.2", "gpt_neox.layers.4", "gpt_neox.layers.6", "gpt_neox.layers.8", "gpt_neox.layers.10"], "layers": [0, 2, 4, 6, 8, 10], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "max_tokens": 1000000000, "log_to_wandb": true, "run_name": "/home/random_sae/experiments/saved_models/pythia-160m_32_k32/redpajama-data-1t-sample_plain_text_100M_trained", "wandb_log_frequency": 1}
gpt_neox.layers.10/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
gpt_neox.layers.10/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1766dfa4afe2cd14c2c149910770cd9e478587f56fc383db198ac589f50aa980
3
+ size 151096648
gpt_neox.layers.2/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
gpt_neox.layers.2/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00cae547405683d18ab68f3105064c8d62d70de9cc9f1177b887839024f30fa1
3
+ size 151096648
gpt_neox.layers.4/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
gpt_neox.layers.4/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad084698c9750bcb1578de3e3459c591abd800a609508a56b76b7673616d708f
3
+ size 151096648
gpt_neox.layers.6/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
gpt_neox.layers.6/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea66c0bfd48f5743fef9e083b7736b8ee5cd7548038a8be5b33c05514ce7e471
3
+ size 151096648
gpt_neox.layers.8/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
gpt_neox.layers.8/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57327847a1bff0a9a648225c6ff16af7b862da885733264192dfb09f18cfe7c8
3
+ size 151096648