< Explain other AI papers

Masking in Multi-hop QA: An Analysis of How Language Models Perform with Context Permutation

Wenyu Huang, Pavlos Vougiouklis, Mirella Lapata, Jeff Z. Pan

2025-05-21

Masking in Multi-hop QA: An Analysis of How Language Models Perform with
  Context Permutation

Summary

This paper talks about how different types of AI models handle tough questions that require using information from several places, and how shuffling the order of that information can affect their answers.

What's the problem?

AI sometimes struggles with multi-hop question answering, which means answering questions that need facts from more than one source, especially when the order of information changes or isn't straightforward.

What's the solution?

The researchers found that encoder-decoder models, which can look at information in both directions, do better at these complex questions than models that only read information one way, especially when the order of the facts is mixed up.

Why it matters?

This matters because it helps us build smarter AI systems that can handle complicated questions more accurately, which is useful for things like research, homework help, and searching for information online.

Abstract

Encoder-decoder models outperform causal decoder-only models in multi-hop question answering by leveraging permutation of search results and bi-directional attention.