Upload 11 files
Browse files- cfg.json +1 -0
- gpt_neox.layers.10/cfg.json +1 -0
- gpt_neox.layers.10/sae.safetensors +3 -0
- gpt_neox.layers.2/cfg.json +1 -0
- gpt_neox.layers.2/sae.safetensors +3 -0
- gpt_neox.layers.4/cfg.json +1 -0
- gpt_neox.layers.4/sae.safetensors +3 -0
- gpt_neox.layers.6/cfg.json +1 -0
- gpt_neox.layers.6/sae.safetensors +3 -0
- gpt_neox.layers.8/cfg.json +1 -0
- gpt_neox.layers.8/sae.safetensors +3 -0
cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false}, "batch_size": 4, "grad_acc_steps": 8, "micro_acc_steps": 2, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["gpt_neox.layers.0", "gpt_neox.layers.2", "gpt_neox.layers.4", "gpt_neox.layers.6", "gpt_neox.layers.8", "gpt_neox.layers.10"], "layers": [0, 2, 4, 6, 8, 10], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "max_tokens": 1000000000, "log_to_wandb": true, "run_name": "/home/random_sae/experiments/saved_models/pythia-160m_32_k32/redpajama-data-1t-sample_plain_text_100M_trained", "wandb_log_frequency": 1}
|
gpt_neox.layers.10/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
|
gpt_neox.layers.10/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1766dfa4afe2cd14c2c149910770cd9e478587f56fc383db198ac589f50aa980
|
3 |
+
size 151096648
|
gpt_neox.layers.2/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
|
gpt_neox.layers.2/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:00cae547405683d18ab68f3105064c8d62d70de9cc9f1177b887839024f30fa1
|
3 |
+
size 151096648
|
gpt_neox.layers.4/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
|
gpt_neox.layers.4/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ad084698c9750bcb1578de3e3459c591abd800a609508a56b76b7673616d708f
|
3 |
+
size 151096648
|
gpt_neox.layers.6/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
|
gpt_neox.layers.6/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ea66c0bfd48f5743fef9e083b7736b8ee5cd7548038a8be5b33c05514ce7e471
|
3 |
+
size 151096648
|
gpt_neox.layers.8/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 768}
|
gpt_neox.layers.8/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:57327847a1bff0a9a648225c6ff16af7b862da885733264192dfb09f18cfe7c8
|
3 |
+
size 151096648
|