Empirical Study on Updating Key-Value Memories in Transformer Feed-forward Layers
2024-02-19Code Available0· sign in to hype
Zihan Qiu, Zeyu Huang, Youcheng Huang, Jie Fu
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/qiuzh20/tuning-keys-v.s.-valuesOfficialIn paperpytorch★ 4
Abstract
The feed-forward networks (FFNs) in transformers are recognized as a group of key-value neural memories to restore abstract high-level knowledge. In this work, we conduct an empirical ablation study on updating keys (the 1st layer in the FFNs layer) or values (the 2nd layer in the FFNs layer). We compare those two methods in various knowledge editing and fine-tuning tasks of large language models to draw insights to understand FFNs further. Code is available at https://github.com/qiuzh20/Tuning-keys-v.s.-valuesthis\,repo.