Update README.md
#1
by
tim-lawson
- opened
README.md
CHANGED
@@ -5,8 +5,14 @@ license: mit
|
|
5 |
tags:
|
6 |
- model_hub_mixin
|
7 |
- pytorch_model_hub_mixin
|
|
|
|
|
8 |
---
|
9 |
|
|
|
|
|
|
|
10 |
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
|
11 |
-
|
12 |
-
-
|
|
|
|
5 |
tags:
|
6 |
- model_hub_mixin
|
7 |
- pytorch_model_hub_mixin
|
8 |
+
datasets:
|
9 |
+
- monology/pile-uncopyrighted
|
10 |
---
|
11 |
|
12 |
+
A Multi-Layer Sparse Autoencoder (MLSAE) trained on [EleutherAI/pythia-70m-deduped](https://huggingface.co/EleutherAI/pythia-70m-deduped) and [monology/pile-uncopyrighted](https://huggingface.co/datasets/monology/pile-uncopyrighted), with an expansion factor of 1 and $k = 16$.
|
13 |
+
For more details, see the [paper](https://arxiv.org/submit/5837813) and the [Weights & Biases project](https://wandb.ai/timlawson-/mlsae).
|
14 |
+
|
15 |
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
|
16 |
+
|
17 |
+
- Library: <https://github.com/tim-lawson/mlsae>
|
18 |
+
- Docs: <https://github.com/tim-lawson/mlsae/blob/main/mlsae/model/autoencoder.py>
|