English
norabelrose commited on
Commit
b860b74
1 Parent(s): 7b95be9

Upload folder using huggingface_hub

Browse files
layers.0.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.0.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae7a57066883a3ab0a5c69619ac22f3c2618d5e667eff42e19a7661b8bba1c42
3
+ size 134351176
layers.0.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.0.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38ed316ac4be8b24a598712f4755c8fcf7514aaec09699cfdd368f54ace417bb
3
+ size 134351176
layers.0/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.0/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b9b454cbdc67614381a9f63a71884823b29fdfbad9c31e1d7da789ce1c18aa0
3
+ size 134351176
layers.1.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.1.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6b6001c0719982f670eea76c163ad476ac74e760245b4e30bd0fc874183f8ab
3
+ size 134351176
layers.1.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.1.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74ec4d0b76c8a3a32bc7e74f1d7da325ce01bde4748c673698cdfbd3b6cce15d
3
+ size 134351176
layers.1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.1/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:557b20eefb2ab497bdf59fabdffdb59e98a614131a4dff6a7209a273aa6bf2d0
3
+ size 134351176
layers.2.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.2.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c16071e1a21646b39ec27758cc83a6736e71658a3922ea91978280150e5d077
3
+ size 134351176
layers.2.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.2.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ccc65d99559aeedd980530120cf32e7643b0185bbb6ae7af5c1a3558d1c080b
3
+ size 134351176
layers.2/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.2/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05dceb38b4081a3bb8ac45b57248bf9f37b2c23dd55e17234d866e12a174d279
3
+ size 134351176
layers.3.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.3.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7766d1a8fd007123c58ef916ac2df6e0f86c893d4fe63fc73e3e4f0baca70509
3
+ size 134351176
layers.3.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.3.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e594ff671d87a7be64416f67fa07632c2fabe7acb355a848bcea7be053f8e2f
3
+ size 134351176
layers.3/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.3/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e536efc671eb55929f42704b72d3bad7b67c906805747bfc4fdc12331f1d596
3
+ size 134351176
layers.4.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.4.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6129edd731fe67d324b215e16e73dc09cd47370497cb6125c05acb8b452fa69c
3
+ size 134351176
layers.4.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.4.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b8deb21b262637a4389557e80a9d980282c6eda4278203212d536fdfde63b88
3
+ size 134351176
layers.4/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.4/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6469f32ad73729a0b8ceb6d4a415d676b93922f5ceead462e67fcf227dd1ba9
3
+ size 134351176
layers.5.attention/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.5.attention/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ee685832631de8aa46b58c3b5d90aeac8aea3c4e7f9870f73d5f5fe471cf4aa
3
+ size 134351176
layers.5.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.5.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e28875f0f851130323c5f48f290d7a95a08d75bf5c5586db4bbcaef878d76c0d
3
+ size 134351176
layers.5/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 32768, "k": 16, "signed": false, "d_in": 512}
layers.5/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68fb4f61bbf9edfb3246ce04d1fd1d0a72c6b8510541ef51962ef4b7d42b1581
3
+ size 134351176