Skip to main content
Journal cover image

Time-Aware Language Models as Temporal Knowledge Bases

Publication ,  Journal Article
Dhingra, B; Cole, JR; Eisenschlos, JM; Gillick, D; Eisenstein, J; Cohen, WW
Published in: Transactions of the Association for Computational Linguistics
March 18, 2022

Many facts come with an expiration date, from the name of the President to the basketball team Lebron James plays for. However, most language models (LMs) are trained on snapshots of data collected at a specific moment in time. This can limit their utility, especially in the closed-book setting where the pretraining corpus must contain the facts the model should memorize. We introduce a diagnostic dataset aimed at probing LMs for factual knowledge that changes over time and highlight problems with LMs at either end of the spectrum—those trained on specific slices of temporal data, as well as those trained on a wide range of temporal data. To mitigate these problems, we propose a simple technique for jointly modeling text with its timestamp. This improves memorization of seen facts from the training time period, as well as calibration on predictions about unseen facts from future time periods. We also show that models trained with temporal context can be efficiently “refreshed” as new data arrives, without the need for retraining from scratch.

Duke Scholars

Altmetric Attention Stats
Dimensions Citation Stats

Published In

Transactions of the Association for Computational Linguistics

DOI

EISSN

2307-387X

Publication Date

March 18, 2022

Volume

10

Start / End Page

257 / 273

Publisher

MIT Press - Journals

Related Subject Headings

  • 4704 Linguistics
  • 4602 Artificial intelligence
  • 2004 Linguistics
  • 1702 Cognitive Sciences
  • 0801 Artificial Intelligence and Image Processing
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Dhingra, B., Cole, J. R., Eisenschlos, J. M., Gillick, D., Eisenstein, J., & Cohen, W. W. (2022). Time-Aware Language Models as Temporal Knowledge Bases. Transactions of the Association for Computational Linguistics, 10, 257–273. https://doi.org/10.1162/tacl_a_00459
Dhingra, Bhuwan, Jeremy R. Cole, Julian Martin Eisenschlos, Daniel Gillick, Jacob Eisenstein, and William W. Cohen. “Time-Aware Language Models as Temporal Knowledge Bases.” Transactions of the Association for Computational Linguistics 10 (March 18, 2022): 257–73. https://doi.org/10.1162/tacl_a_00459.
Dhingra B, Cole JR, Eisenschlos JM, Gillick D, Eisenstein J, Cohen WW. Time-Aware Language Models as Temporal Knowledge Bases. Transactions of the Association for Computational Linguistics. 2022 Mar 18;10:257–73.
Dhingra, Bhuwan, et al. “Time-Aware Language Models as Temporal Knowledge Bases.” Transactions of the Association for Computational Linguistics, vol. 10, MIT Press - Journals, Mar. 2022, pp. 257–73. Crossref, doi:10.1162/tacl_a_00459.
Dhingra B, Cole JR, Eisenschlos JM, Gillick D, Eisenstein J, Cohen WW. Time-Aware Language Models as Temporal Knowledge Bases. Transactions of the Association for Computational Linguistics. MIT Press - Journals; 2022 Mar 18;10:257–273.
Journal cover image

Published In

Transactions of the Association for Computational Linguistics

DOI

EISSN

2307-387X

Publication Date

March 18, 2022

Volume

10

Start / End Page

257 / 273

Publisher

MIT Press - Journals

Related Subject Headings

  • 4704 Linguistics
  • 4602 Artificial intelligence
  • 2004 Linguistics
  • 1702 Cognitive Sciences
  • 0801 Artificial Intelligence and Image Processing