Information Association for Language Model Updating by Mitigating LM-Logical Discrepancy

Pengfei Yu, Heng Ji

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Large Language Models (LLMs) struggle with providing current information due to the outdated pre-training data. Existing methods for updating LLMs, such as knowledge editing and continual fine-tuning, have significant drawbacks in generalizability of new information and the requirements on structured updating corpus. We identify the core challenge behind these drawbacks: the LM-logical discrepancy featuring the difference between language modeling probabilities and logical probabilities. To evaluate and address the core challenge, we propose a new task formulation of the information updating task that only requires the provision of an unstructured updating corpus and evaluates the performance of information updating on the generalizability to question-answer pairs pertaining to the updating information. We further propose a novel and effective pipeline approach for the task, highlighting a self-prompting-based question-answer generation process and a associative distillation methods to bridge the LM-logical discrepancy. We develop two datasets for evaluation, one sourced from news articles published in March and April 20231, and the other from the Natural Questions benchmark. Experimental results demonstrate the superiority of our approach, significantly increasing the factual consistency score (on a scale from 0 to 1) by up to 0.16. Furthermore, our method effectively mitigates forgetting utilizing a compact replay buffer with only 2.3% of the training tokens.

Original languageEnglish (US)
Title of host publicationCoNLL 2024 - 28th Conference on Computational Natural Language Learning, Proceedings of the Conference
EditorsLibby Barak, Malihe Alikhani
PublisherAssociation for Computational Linguistics (ACL)
Pages117-129
Number of pages13
ISBN (Electronic)9798891761780
DOIs
StatePublished - 2024
Event28th Conference on Computational Natural Language Learning, CoNLL 2024 - Miami, United States
Duration: Nov 15 2024Nov 16 2024

Publication series

NameCoNLL 2024 - 28th Conference on Computational Natural Language Learning, Proceedings of the Conference

Conference

Conference28th Conference on Computational Natural Language Learning, CoNLL 2024
Country/TerritoryUnited States
CityMiami
Period11/15/2411/16/24

ASJC Scopus subject areas

  • Artificial Intelligence
  • Human-Computer Interaction
  • Linguistics and Language

Fingerprint

Dive into the research topics of 'Information Association for Language Model Updating by Mitigating LM-Logical Discrepancy'. Together they form a unique fingerprint.

Cite this