Bidirectional LMs are Better Knowledge Memorizers? A Benchmark for Real-world Knowledge Injection
Yuwei Zhang, Wenhao Yu, Shangbin Feng, Yifan Zhu, Letian Peng, Jayanth Srinivasa, Gaowen Liu, Jingbo Shang
2025-05-21
Summary
This paper talks about a new way to test how well different types of AI language models remember and use real-world facts by using a special benchmark called WikiDYK.
What's the problem?
AI models sometimes struggle to accurately remember and use information, and it's not clear which type of model does this best or how to improve their reliability.
What's the solution?
The researchers created the WikiDYK benchmark to compare how well two types of language models, CLMs and BiLMs, memorize knowledge, and they introduced a new system where different parts of the AI can work together to make the model more dependable.
Why it matters?
This matters because it helps scientists and developers figure out how to build AI that remembers facts more accurately, making these models more useful and trustworthy for real-world tasks.
Abstract
A new knowledge injection benchmark, WikiDYK, highlights differences in knowledge memorization between CLMs and BiLMs, and a modular collaborative framework enhances LLM reliability.