Upload folder using huggingface_hub
Browse files- config.json +1 -0
- layers.0.mlp/cfg.json +1 -0
- layers.0.mlp/sae.safetensors +3 -0
- layers.1.mlp/cfg.json +1 -0
- layers.1.mlp/sae.safetensors +3 -0
- layers.10.mlp/cfg.json +1 -0
- layers.10.mlp/sae.safetensors +3 -0
- layers.11.mlp/cfg.json +1 -0
- layers.11.mlp/sae.safetensors +3 -0
- layers.12.mlp/cfg.json +1 -0
- layers.12.mlp/sae.safetensors +3 -0
- layers.13.mlp/cfg.json +1 -0
- layers.13.mlp/sae.safetensors +3 -0
- layers.14.mlp/cfg.json +1 -0
- layers.14.mlp/sae.safetensors +3 -0
- layers.15.mlp/cfg.json +1 -0
- layers.15.mlp/sae.safetensors +3 -0
- layers.2.mlp/cfg.json +1 -0
- layers.2.mlp/sae.safetensors +3 -0
- layers.3.mlp/cfg.json +1 -0
- layers.3.mlp/sae.safetensors +3 -0
- layers.4.mlp/cfg.json +1 -0
- layers.4.mlp/sae.safetensors +3 -0
- layers.5.mlp/cfg.json +1 -0
- layers.5.mlp/sae.safetensors +3 -0
- layers.6.mlp/cfg.json +1 -0
- layers.6.mlp/sae.safetensors +3 -0
- layers.7.mlp/cfg.json +1 -0
- layers.7.mlp/sae.safetensors +3 -0
- layers.8.mlp/cfg.json +1 -0
- layers.8.mlp/sae.safetensors +3 -0
- layers.9.mlp/cfg.json +1 -0
- layers.9.mlp/sae.safetensors +3 -0
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true}, "batch_size": 4, "grad_acc_steps": 2, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["layers.0.mlp", "layers.1.mlp", "layers.2.mlp", "layers.3.mlp", "layers.4.mlp", "layers.5.mlp", "layers.6.mlp", "layers.7.mlp", "layers.8.mlp", "layers.9.mlp", "layers.10.mlp", "layers.11.mlp", "layers.12.mlp", "layers.13.mlp", "layers.14.mlp", "layers.15.mlp"], "init_seeds": [0], "layers": [], "layer_stride": 1, "transcode": true, "distribute_modules": true, "save_every": 1000, "log_to_wandb": true, "run_name": "llama-1b-transcode", "wandb_log_frequency": 1, "model": "meta-llama/Llama-3.2-1B", "dataset": "EleutherAI/rpj-v2-sample", "split": "train", "ctx_len": 2048, "revision": null, "load_in_8bit": false, "max_examples": null, "resume": false, "text_column": "raw_content", "finetune": null, "shuffle_seed": 42, "data_preprocessing_num_proc": 112}
|
layers.0.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.0.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c5eae597555b4879aa1dd37d1135c6bb41f5977089ae307a8d8f278596affb6d
|
3 |
+
size 2164793776
|
layers.1.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.1.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc61ee9e3cce077aafd611546c28f10c6cf780310869f91923c73f144cba616d
|
3 |
+
size 2164793776
|
layers.10.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.10.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28de292afccb493a6ac69f89e8282b10f8ae79eb3388c46305490ccf4c1e55e4
|
3 |
+
size 2164793776
|
layers.11.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.11.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39cd3ea9555dc71d366815b6c3c157d4b4f0b5f5090b6fdbc129782ebccde92f
|
3 |
+
size 2164793776
|
layers.12.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.12.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e05cc08f1831962aab7d71bc993a64f8b14f7230640fa542ac003b87b3e96189
|
3 |
+
size 2164793776
|
layers.13.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.13.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d5f69bfa0cfb239b14985d39504f5385eb6547b534ce3a7aa2d7d21868462f5d
|
3 |
+
size 2164793776
|
layers.14.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.14.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7eb66dd466acbfe828e42f7dcf1dc538aebc85005b794b172d38bfef41af9467
|
3 |
+
size 2164793776
|
layers.15.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.15.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83ca3e1ab63ad2476dcc42f008088bef2e7cdd86b9217a43c69f62c2cbf41c73
|
3 |
+
size 2164793776
|
layers.2.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.2.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39902e807ea3e183a1aa1cc1f3a633ca427dea0c6a241dd20a1f91c1059d6075
|
3 |
+
size 2164793776
|
layers.3.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.3.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff16d20711a7ca72f36230a0967fd88a8e1ef9eee586c4a5b80cc288115eb5be
|
3 |
+
size 2164793776
|
layers.4.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.4.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0735e73bcb30f3f5021b197ef8e17f391773555933debac3d3abb0e8ab4cfe4f
|
3 |
+
size 2164793776
|
layers.5.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.5.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f75873601365792e7ff2c4921fbd047789916a6dbdfb595ffbf8b9305e8cc9f7
|
3 |
+
size 2164793776
|
layers.6.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.6.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2443554465d9b51e0d22ded5eba06feb60e656ea9250cd6fc55d54777748368c
|
3 |
+
size 2164793776
|
layers.7.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.7.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd33909a7de06d1a443028bbcb9dd926a249367c72b99df80b373281cfd49a60
|
3 |
+
size 2164793776
|
layers.8.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.8.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8f4e410c59977785f3da28930b10eb2120e3a58cbca92ea09c54d024797b02ef
|
3 |
+
size 2164793776
|
layers.9.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 32, "multi_topk": false, "skip_connection": true, "d_in": 2048}
|
layers.9.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee06e6f63dbf738b6901f4bee961e9f36c878bc11c070407dcf68c8c2b690427
|
3 |
+
size 2164793776
|