Abstract

Schema matching over relational data has been studied for more than two decades. However, the state-of-the-art methods do not address key modern-day challenges encountered in real customer scenarios, namely: 1) no access to the source (customer) data due to privacy constraints, 2) target schema with a much larger number of entities and attributes compared to the source schema, and 3) different but semantically equivalent entity and attribute names in the source and target schemata. In this paper, we address these shortcomings. Using real-world customer schemata, we demonstrate that existing linguistic matching approaches have low accuracy. Next, we propose the Learned Schema Mapper (LSM), a novel linguistic schema matching system that leverages the natural language understanding capabilities of pre-trained language models to improve the overall accuracy. Combining this with active learning and a smart attribute selection strategy that selects the most informative attributes for users to label, LSM can significantly reduce the overall human labeling cost. Experimental results demonstrate that users can correctly match their full schema while saving as much as 81% of the labeling cost compared to manual labeling.

Keywords:
Computer science Schema matching Schema (genetic algorithms) Database schema Artificial intelligence Natural language processing Matching (statistics) Information retrieval Natural language Machine learning Data mining Data integration Database design

Metrics

18
Cited By
4.60
FWCI (Field Weighted Citation Impact)
45
Refs
0.94
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Text and Document Classification Technologies
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Schema matching based on energy domain pre-trained language model

Zhiyu PanMuchen YangAntonello Monti

Journal:   Energy Informatics Year: 2023 Vol: 6 (S1)
JOURNAL ARTICLE

Deep entity matching with pre-trained language models

Yuliang LiJinfeng LiYoshihiko SuharaAnHai DoanWang-Chiew Tan

Journal:   Proceedings of the VLDB Endowment Year: 2020 Vol: 14 (1)Pages: 50-60
BOOK-CHAPTER

Pre-trained Language Models

Huaping ZhangJianyun Shang

Year: 2025 Pages: 73-90
BOOK-CHAPTER

Pre-trained Language Models

Gerhard PaaßSven Giesselbach

Artificial intelligence: foundations, theory, and algorithms/Artificial intelligence: Foundations, theory, and algorithms Year: 2023 Pages: 19-78
© 2026 ScienceGate Book Chapters — All rights reserved.