Multi-Layer Sparse Autoencoders with Transformers
Collection
Single SAEs trained on the residual stream activation vectors from every transformer layer simultaneously (including the transformers).
•
30 items
•
Updated
This model has been pushed to the Hub using the PytorchModelHubMixin integration: