We believe that the parametric knowledge in LLMs is still a largely unexplored area, and we hope this repository will provide you with some valuable insights!😶🌫️
-
Decoding specialised feature neurons in LLMs with the final projection layer
[Logits Lens, Analysis of Query Neuron]
-
What does the knowledge neuron thesis have to do with knowledge?
Jingcheng Niu, Andrew Liu, Zining Zhu, Gerald Penn. ICLR'24(Spotlight)
-
Knowledge Mechanisms in Large Language Models: A Survey and Perspective
Mengru Wang, Yunzhi Yao, Ziwen Xu, Shuofei Qiao, Shumin Deng, Peng Wang, Xiang Chen, Jia-Chen Gu, Yong Jiang, Pengjun Xie, Fei Huang, Huajun Chen, Ningyu Zhang. EMNLP'24 Findings
-
Disentangling Memory and Reasoning Ability in Large Language Models
Mingyu Jin, Weidi Luo, Sitao Cheng, Xinyi Wang, Wenyue Hua, Ruixiang Tang, William Yang Wang, Yongfeng Zhang. preprint'24
-
Linguistic collapse: Neural collapse in (large) language models
Robert Wu, Vardan Papyan. NIPS'24
-
Understanding the Interplay between Parametric and Contextual Knowledge for Large Language Models
Sitao Cheng, Liangming Pan, Xunjian Yin, Xinyi Wang, William Yang Wang. Preprint'24
-
Evaluating the External and Parametric Knowledge Fusion of Large Language Models
Hao Zhang, Yuyang Zhang, Xiaoguang Li, Wenxuan Shi, Haonan Xu, Huanshuo Liu, Yasheng Wang, Lifeng Shang, Qun Liu, Yong Liu, Ruiming Tang. Preprint'24
-
Adaptive Chameleon or Stubborn Sloth: Revealing the Behavior of Large Language Models in Knowledge Conflicts
Jian Xie, Kai Zhang, Jiangjie Chen, Renze Lou, Yu Su. ICLR'24 Spotlight
-
Knowledge entropy decay during language model pretraining hinders new knowledge acquisition
Jiyeon Kim, Hyunji Lee, Hyowon Cho, Joel Jang, Hyeonbin Hwang, Seungpil Won, Youbin Ahn, Dohaeng Lee, Minjoon Seo. Preprint'24
-
When Context Leads but Parametric Memory Follows in Large Language Models
Yufei Tao, Adam Hiatt, Erik Haake, Antonie J. Jetter, Ameeta Agrawal. EMNLP'24
-
Neuron-level knowledge attribution in large language models
Zeping Yu, Sophia Ananiadou. EMNLP'24
-
Dissecting recall of factual associations in auto-regressive language models[code]
Mor Geva, Jasmijn Bastings, Katja Filippova, Amir Globerson. EMNLP'23
-
Transformer Feed-Forward Layers Are Key-Value Memories
Mor Geva, Roei Schuster, Jonathan Berant, Omer Levy. EMNLP'21
-
Yifan Wei, Xiaoyan Yu, Yixuan Weng, Huanhuan Ma, Yuanzhe Zhang, Jun Zhao, Kang Liu. CIKM'24
-
Locating and Editing Factual Associations in GPT
Kevin Meng, David Bau, Alex Andonian, Yonatan Belinkov. NIPS'22
-
Identifying query-relevant neurons in large language models for long-form texts
Lihu Chen, Adam Dejl, Francesca Toni. Preprint'24
-
Revealing the parametric knowledge of language models: A unified framework for attribution methods
Haeun Yu, Pepa Atanasova, Isabelle Augenstein. ACL'24
-
Does Large Language Model contain Task-Specific Neurons.
Ran Song, Shizhu He, Shuting Jiang, Yantuan Xian, Shengxiang Gao, Kang Liu, and Zhengtao Yu. EMNLP'24
-
Journey to the center of the knowledge neurons: Discoveries of language-independent knowledge neurons and degenerate knowledge neurons
Yuheng Chen, Pengfei Cao, Yubo Chen, Kang Liu, Jun Zhao. AAAI'24
-
Knowledge Neurons in Pretrained Transformers
Damai Dai, Li Dong, Yaru Hao, Zhifang Sui, Baobao Chang, Furu Wei. ACL'22
-
Separating tongue from thought: Activation patching reveals language-agnostic concept representations in transformers
Clément Dumas, Chris Wendler, Veniamin Veselovsky, Giovanni Monea, Robert West. ICLR'24 Spotlight
-
From yes-men to truth-tellers Addressing sycophancy in large language models with pinpoint tuning
Wei Chen, Zhen Huang, Liang Xie, Binbin Lin, Houqiang Li, Le Lu, Xinmei Tian, Deng Cai, Yonggang Zhang, Wenxiao Wang, Xu Shen, Jieping Ye. ICML'24
-
Language-specific neurons: The key to multilingual capabilities in large language models.
Tianyi Tang, Wenyang Luo, Haoyang Huang, Dongdong Zhang, Xiaolei Wang, Xin Zhao, Furu Wei, Ji-Rong Wen. ACL'24
-
Multi-property Steering of Large Language Models with Dynamic Activation Composition
Daniel Scalena, Gabriele Sarti, Malvina Nissim. ACL'24 BlackboxNLP Workshop
-
Exploring the benefit of activation sparsity in pre-training
[MoE, Activation Sparsity, Activation Pattern, Inference Speedup] Zhengyan Zhang, Chaojun Xiao, Qiujieli Qin, Yankai Lin, Zhiyuan Zeng, Xu Han, Zhiyuan Liu, Ruobing Xie, Maosong Sun, Jie Zhou. ICML'24
-
Activation Addition: Steering Language Models Without Optimization
Alexander Matt Turner, Lisa Thiergart, Gavin Leech, David Udell, Juan J. Vazquez, Ulisse Mini, Monte MacDiarmid. Preprint'23
-
Deja vu: Contextual sparsity for efficient LLMs at inference time
[Sparsity, Inference Speedup] Zichang Liu, Jue Wang, Tri Dao, Tianyi Zhou, Binhang Yuan, Zhao Song, Anshumali Shrivastava, Ce Zhang, Yuandong Tian, Christopher Re, Beidi Chen. ICML'23
-
A Comprehensive Study of Knowledge Editing for Large Language Models
Ningyu Zhang, Yunzhi Yao, Bozhong Tian, Peng Wang, Shumin Deng, Mengru Wang, Zekun Xi, Shengyu Mao, Jintian Zhang, Yuansheng Ni, Siyuan Cheng, Ziwen Xu, Xin Xu, Jia-Chen Gu, Yong Jiang, Pengjun Xie, Fei Huang, Lei Liang, Zhiqiang Zhang, Xiaowei Zhu, Jun Zhou, Huajun Chen. Preprint'24
-
FAME: Towards Factual Multi-Task Model Editing Li Zeng, Yingyu Shan, Zeming Liu, Jiashu Yao, Yuhang Guo. EMNLP'24
-
To Forget or Not? Towards Practical Knowledge Unlearning for Large Language Models
Bozhong Tian, Xiaozhuan Liang, Siyuan Cheng, Qingbin Liu, Mengru Wang, Dianbo Sui, Xi Chen, Huajun Chen, Ningyu Zhang. EMNLP'24 findings
-
Understanding the Collapse of LLMs in Model Editing
Wanli Yang, Fei Sun, Jiajun Tan, Xinyu Ma, Du Su, Dawei Yin, Huawei Shen. EMNLP'24 findings
-
Is it possible to edit large language models robustly?
Xinbei Ma, Tianjie Ju, Jiyang Qiu, Zhuosheng Zhang, Hai Zhao, Lifeng Liu, Yulong Wang. Preprint'24
-
Retrieval-enhanced Knowledge Editing in Language Models for Multi-Hop Question Answering
Yucheng Shi, Qiaoyu Tan, Xuansheng Wu, Shaochen Zhong, Kaixiong Zhou, Ninghao Liu. CIKM'24
-
Latent paraphrasing: Perturbation on layers improves knowledge injection in language models
Minki Kang, Sung Ju Hwang, Gibbeum Lee, Jaewoong Cho. NIPS'24
-
Learning to edit: Aligning LLMs with knowledge editing
Yuxin Jiang, Yufei Wang, Chuhan Wu, Wanjun Zhong, Xingshan Zeng, Jiahui Gao, Liangyou Li, Xin Jiang, Lifeng Shang, Ruiming Tang, Qun Liu, Wei Wang. ACL'24
-
Inspecting and Editing Knowledge Representations in Language Models
Evan Hernandez, Belinda Z. Li, Jacob Andreas. COLM'24
-
Shiwen Ni, Dingwei Chen, Chengming Li, Xiping Hu, Ruifeng Xu, Min Yang. ACL'24
-
Ethos: Rectifying language models in orthogonal parameter space
[Toxic/Bias Unlearning, SVD, Analysis of Parametric Knowledge, Task Vector]
NAACL'24 findings
-
Editing Large Language Models: Problems, Methods, and Opportunities
Yunzhi Yao, Peng Wang, Bozhong Tian, Siyuan Cheng, Zhoubo Li, Shumin Deng, Huajun Chen, Ningyu Zhang. EMNLP'23
-
Locating and Editing Factual Associations in GPT
Kevin Meng, David Bau, Alex Andonian, Yonatan Belinkov. NIPS'22
-
Memory-Based Model Editing at Scale
Eric Mitchell, Charles Lin, Antoine Bosselut, Christopher D. Manning, Chelsea Finn. ICLR'22
-
Editing Factual Knowledge in Language Models
Nicola De Cao, Wilker Aziz, Ivan Titov. EMNLP'21
-
Anton Sinitsin, Vsevolod Plokhotnyuk, Dmitriy Pyrkin, Sergei Popov, Artem Babenko. ICLR'20
Ming Zhong, Chenxin An, Weizhu Chen, Jiawei Han, Pengcheng He. ICLR'24
-
Initializing models with larger ones
Zhiqiu Xu, Yanjie Chen, Kirill Vishniakov, Yida Yin, Zhiqiang Shen, Trevor Darrell, Lingjie Liu, Zhuang Liu. ICLR'24 Spotlight
-
Cross-model Control: Improving Multiple Large Language Models in One-time Training
Jiayi Wu, Hao Sun, Hengyi Cai, Lixin Su, Shuaiqiang Wang, Dawei Yin, Xiang Li, Ming Gao. NIPS'24
-
Knowledge fusion of large language models
Fanqi Wan, Xinting Huang, Deng Cai, Xiaojun Quan, Wei Bi, Shuming Shi. ICLR'24
-
Tuning language models by proxy
Alisa Liu, Xiaochuang Han, Yizhong Wang, Yulia Tsvetkov, Yejin Choi, Noah A. Smith. COLM'24
-
Chat vector: A simple approach to equip LLMs with instruction following and model alignment in new languages
[Task Vector, Parametric Knowledge, Knowledge Transfer]
ACL'24
-
FedMKT: Federated Mutual Knowledge Transfer for Large and Small Language Models
[Federated Learning, Knowledge Transfer, Heterogeneous Token Alignment]
Coling'25
-
Function vectors in large language models
[Function Vector, Causal Mediation, Mechanism Interpretation]
ICLR'24
-
Refine large language model fine-tuning via instruction vector
[Catastrophic Forgetting, Function Vector, Causal Mediation]
Preprint'24
-
KlF: Knowledge localization and fusion for language model continual learning
[Catastrophic Forgetting, Continual Learning, Sensetity-based Location]
ACL'24
-
Language models are super mario: Absorbing abilities from homologous models as a free lunch
[Knowledge Transfer, Model Merging, Efficient Skill] ICML'24
-
Beyond task vectors: Selective task arithmetic based on importance metrics
[Task Vector, Sensetivity-based Importance Score, Model Merging] Preprint'24
-
Mutual enhancement of large and small language models with cross-silo knowledge transfer
Yongheng Deng, Ziqing Qiao, Ju Ren, Yang Liu, Yaoxue Zhang. Preprint'23
-
Learning to grow pretrained models for efficient transformer training
Peihao Wang, Rameswar Panda, Lucas Torroba Hennigen, Philip Greengard, Leonid Karlinsky, Rogerio Feris, David D. Cox, Zhangyang Wang, Yoon Kim. ICLR'23
-
Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Ran Lucien Wang, Rui Yan. EMNLP'23 Findings
-
Editing models with task arithmetic
[Task Vecotr, Parametric Knowledge, Knowledge Transfer, Multi-task Learning]
ICLR'23
-
Task-Specific Skill Localization in Fine-tuned Language Models
[Knowledge Transfer, Model Graft, Skill Parameter Localization]
ICML'23
-
Composing parameter-efficient modules with arithmetic operations
[PEFT, Task Vector, Model Merge]
NIPS'23
-
Dataless knowledge fusion by merging weights of language models
[Model Merge]
ICLR'23
-
Weight distillation: Transferring the knowledge in neural network parameters
Ye Lin, Yanyang Li, Ziyang Wang, Bei Li, Quan Du, Tong Xiao, Jingbo Zhu. ACL'21
-
Multi-property Steering of Large Language Models with Dynamic Activation Composition
Daniel Scalena, Gabriele Sarti, Malvina Nissim. ACL'24 BlackboxNLP Workshop
-
Word embeddings are steers for language models
[Word Embedding Steering, Generation Control] ACL'24
Alexander Matt Turner, Lisa Thiergart, Gavin Leech, David Udell, Juan J. Vazquez, Ulisse Mini, Monte MacDiarmid. Preprint'23
-
PromptKD: Distilling Student-Friendly Knowledge for Generative Language Models via Prompt Tuning(Note: not parametric)
Gyeongman Kim, Doohyuk Jang, Eunho Yang. EMNLP'24 findings
-
From Instance Training to Instruction Learning: Task Adapters Generation from Instructions
Huanxuan Liao, Yao Xu, Shizhu He, Yuanzhe Zhang, Yanchao Hao, Shengping Liu, Kang Liu, Jun Zhao. NIPS'24
-
Srikrishna Iyer. EMNLP'24 CoNLL Workshop
-
OneBit: Towards extremely low-bit large language models
Yuzhuang Xu, Xu Han, Zonghan Yang, Shuo Wang, Qingfu Zhu, Zhiyuan Liu, Weidong Liu, Wanxiang Che. NIPS'24
-
The cost of compression: Investigating the impact of compression on parametric knowledge in language models
Satya Sai Srinath Namburi, Makesh Sreedhar, Srinath Srinivasan, Frederic Sala. EMNLP'23 findings
-
Awakening augmented generation: Learning to awaken internal knowledge of large language models for question answering
[HyperNet, RAG, Context Compression]
Huanxuan Liao, Shizhu He, Yao Xu, Yuanzhe Zhang, Kang Liu, Shengping Liu, Jun Zhao. AAAI'25
-
Memory injections: Correcting multi-hop reasoning failures during inference in transformer-based language models
Mansi Sakarvadia, Aswathy Ajith, Arham Khan, Daniel Grzenda, Nathaniel Hudson, André Bauer, Kyle Chard, Ian Foster. Oral Presentation at BlackboxNLP Workshop at EMNLP'23
-
Decouple knowledge from parameters for plug-and-play language modeling
Xin Cheng, Yankai Lin, Xiuying Chen, Dongyan Zhao, Rui Yan. ACL'23 findings
-
IN-PARAMETER KNOWLEDGE INJECTION: INTEGRATING TEMPORARY CONTEXTUAL INFORMATION INTO MODEL PARAMETERS
submitted to ICLR'25
-
Kformer: Knowledge injection in transformer feed-forward layers
Yunzhi Yao, Shaohan Huang, Li Dong, Furu Wei, Huajun Chen, Ningyu Zhang. NLPCC'22
-
KaSA: Knowledge-aware singular-value adaptation of large language models
[Knowledge-aware LoRA, SVD]
Fan Wang, Juyong Jiang, Chansung Park, Sunghun Kim, Jing Tang. Preprint'24
-
CorDA: Context-Oriented Decomposition Adaptation of Large Language Models for Task-Aware Parameter-Efficient Fine-tuning
[Knowledge-aware LoRA, SVD]
Yibo Yang, Xiaojie Li, Zhongzhu Zhou, Shuaiwen Leon Song, Jianlong Wu, Liqiang Nie, Bernard Ghanem. NIPS'24
-
DoRA: Weight-Decomposed Low-Rank Adaptation
[Weight-Decomposed LoRA, SVD, Analysis of FT and LoRA] Shih-Yang Liu, Chien-Yi Wang, Hongxu Yin, Pavlo Molchanov, Yu-Chiang Frank Wang, Kwang-Ting Cheng, Min-Hung Chen. ICML'24 Oral
-
Low-rank adaptation with task-relevant feature enhancement for fine-tuning language models
[Task-aware LoRA, Hidden Representation Enhancement] AAAI'25 CoLoRAI Workshop
-
Learn more, but bother less: Parameter efficient continual learning
[Continual Learning, Parameter Efficient, Knowledge Transfer] NIPS'24
-
What will my model forget? Forecasting forgotten examples in language model refinement
[Catastrophic Forgetting, Forecasting Forgetting, Analysis] ICML'24 Spotlight
-
xRAG: Extreme context compression for retrieval-augmented generation with one token
[Context Compression, RAG, Multimodal Fusion] NIPS'24
-
LongEmbed: Extending embedding models for long context retrieval
[Long Context, Embedding Model, Benchmark] EMNLP'24
-
LLM maybe LongLM: Self-extend LLM context window without tuning
[Long Context Extend, Plug-and-Play Method] ICML'24 Spotlight
-
Two stones hit one bird: Bilevel positional encoding for better length extrapolation
[Long Context Extend, Absolute PE + Relative PE, Plug-and-Play but Training-based Method] ICML'24
-
YaRN: Efficient context window extension of large language models[http://arxiv.org/abs/2309.00071]
[Long Context Extend, Variation of RoPE] ICLR'24
-
Train short, test long: Attention with linear biases enables input length extrapolation
[Alibi, Long Context Extrapolate, Training-based Method] ICLR'22
-
RoFormer: Enhanced Transformer with Rotary Position Embedding.
[Rotary Position Embedding, Classic]