JOURNAL ARTICLE

Birds have four legs?! NumerSense: Probing Numerical Commonsense Knowledge of Pre-Trained Language Models

Abstract

Recent works show that pre-trained language models (PTLMs), such as BERT, possess certain commonsense and factual knowledge. They suggest that it is promising to use PTLMs as “neural knowledge bases” via predicting masked words. Surprisingly, we find that this may not work for numerical commonsense knowledge (e.g., a bird usually has two legs). In this paper, we investigate whether and to what extent we can induce numerical commonsense knowledge from PTLMs as well as the robustness of this process. In this paper, we investigate whether and to what extent we can induce numerical commonsense knowledge from PTLMs as well as the robustness of this process. To study this, we introduce a novel probing task with a diagnostic dataset, NumerSense, containing 13.6k masked-word-prediction probes (10.5k for fine-tuning and 3.1k for testing). Our analysis reveals that: (1) BERT and its stronger variant RoBERTa perform poorly on the diagnostic dataset prior to any fine-tuning; (2) fine-tuning with distant supervision brings some improvement; (3) the best supervised model still performs poorly as compared to human performance (54.06% vs. 96.3% in accuracy).

Keywords:
Commonsense knowledge Robustness (evolution) Computer science Artificial intelligence Natural language processing Task (project management) Word-sense disambiguation Language model Deep neural networks Commonsense reasoning Machine learning Artificial neural network Process (computing) Knowledge extraction

Metrics

114
Cited By
14.10
FWCI (Field Weighted Citation Impact)
32
Refs
0.99
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

Probing Simile Knowledge from Pre-trained Language Models

Weijie ChenYongzhu ChangRongsheng ZhangJiashu PuGuandan ChenLe ZhangYadong XiYijiang ChenChang Su

Journal:   Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) Year: 2022 Pages: 5875-5887
JOURNAL ARTICLE

Evaluating Commonsense in Pre-Trained Language Models

Xuhui ZhouYue ZhangLeyang CuiDandan Huang

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2020 Vol: 34 (05)Pages: 9733-9740
© 2026 ScienceGate Book Chapters — All rights reserved.