--- license: apache-2.0 datasets: - enjalot/fineweb-edu-sample-10BT-chunked-500-nomic-text-v1.5 language: - en --- # Latent SAE A series of SAEs trained on embeddings from [nomic-embed-text-v1.5](https://huggingface.co/nomic-ai/nomic-embed-text-v1.5) The SAEs were trained on the 100BT sample of Fineweb-EDU, see an example of the [10BT sample of Fineweb-Edu](https://huggingface.co/datasets/enjalot/fineweb-edu-sample-10BT-chunked-500). Run the models or train your own with [Latent SAE](https://github.com/enjalot/latent-sae) which is heavily borrowing from https://github.com/EleutherAI/sae # Training The models were trained using Modal Labs infrastructure with the command: ```bash modal run train_modal.py --batch-size 512 --grad-acc-steps 4 --k 64 --expansion-factor 32 ``` Error and dead latents charts can be seen here: ![image/png](https://cdn-uploads.huggingface.co/production/uploads/631bce12bf1351ed2bd6bffe/GKPdI97ogF5tF709oYbbY.png)