Time is Encoded in the Weights of Finetuned Language Models
Abstract
We present time vectors, a simple tool to customize language models to new time periods. Time vectors are created by finetuning a language model on data from a single time (e.g., a year or month), and then subtracting the weights of the original pretrained model. This vector specifies a direction in weight space that, as our experiments show, improves performance on text from that time period. Time vectors specialized to adjacent time periods appear to be positioned closer together in a manifold. Using this structure, we interpolate between time vectors to induce new models that perform better on intervening and future time periods, without any additional training. We demonstrate the consistency of our findings across different tasks, domains, model sizes, and time scales. Our results suggest that time is encoded in the weight space of finetuned models.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Language and Task Arithmetic with Parameter-Efficient Layers for Zero-Shot Summarization (2023)
- Efficiently Adapting Pretrained Language Models To New Languages (2023)
- PEMA: Plug-in External Memory Adaptation for Language Models (2023)
- LM-Cocktail: Resilient Tuning of Language Models via Model Merging (2023)
- Tied-Lora: Enhacing parameter efficiency of LoRA with weight tying (2023)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper