wlog commited on
Commit
eae1a73
·
verified ·
1 Parent(s): aec3970

Upload 13 files

Browse files
cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false}, "batch_size": 4, "grad_acc_steps": 8, "micro_acc_steps": 2, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["gpt_neox.layers.0", "gpt_neox.layers.1", "gpt_neox.layers.2", "gpt_neox.layers.3", "gpt_neox.layers.4", "gpt_neox.layers.5"], "layers": [0, 1, 2, 3, 4, 5], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "max_tokens": 1000000000, "log_to_wandb": true, "run_name": "/home/random_sae/experiments/saved_models/pythia-70m_32_k32/redpajama-data-1t-sample_plain_text_100M_trained", "wandb_log_frequency": 1}
gpt_neox.layers.0/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
gpt_neox.layers.0/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:893c3c56ac75dc78d3d877a996a184d880fd3f4766b1a472001061ac69927ab4
3
+ size 67176776
gpt_neox.layers.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
gpt_neox.layers.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe68197f22f6b1ed40b790766f3e4a0bc7bb9a77ebb7c4a387bd2e73cf4061f6
3
+ size 67176776
gpt_neox.layers.2/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
gpt_neox.layers.2/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:204e1029db40be61d71dd07827e2bc8fdb37e303d8bc07d4bf470282fa366348
3
+ size 67176776
gpt_neox.layers.3/cfg.json CHANGED
@@ -1 +1 @@
1
- {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 16, "multi_topk": false, "d_in": 512}
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
gpt_neox.layers.3/sae.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b996eb7968119c84b0251df25f3819f79151357bd51db4adcaaf6f19c97fe49e
3
  size 67176776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f94098f98579263de0beea3bcf8a2f51b30b1d1cbcb6805f3a2ae74ee19ad477
3
  size 67176776
gpt_neox.layers.4/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
gpt_neox.layers.4/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:458350f5f4d6afa9021f01355506bc5ee03049054c52a7dd7b7978795e983452
3
+ size 67176776
gpt_neox.layers.5/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "d_in": 512}
gpt_neox.layers.5/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ab69ec0ffda1ad2038efafaa442bae9446a058fcc5ab444fc70c4972ab520b9
3
+ size 67176776