
MIT Researchers Uncover Position Bias in Large Language Models
Large language models (LLMs), the engines behind many AI applications, have a hidden flaw: they tend to overemphasize information presented at the beginning and end of documents or conversations, a phenomenon known as “position bias.” This can lead to inconsistencies and inaccuracies, especially in tasks requiring comprehensive analysis.
Researchers at MIT have made a significant breakthrough in understanding the underlying mechanisms that cause this bias. Their work not only pinpoints the origins of the problem but also proposes a framework for diagnosing and correcting it in future model designs.
The team’s research, led by Xinyi Wu, a graduate student at MIT’s Institute for Data, Systems, and Society (IDSS) and the Laboratory for Information and Decision Systems (LIDS), reveals that the architecture of LLMs, particularly how they process input data, plays a crucial role in creating position bias. Additionally, the training data used to develop these models can also contribute to the problem.
“These models are black boxes, so as an LLM user, you probably don’t know that position bias can cause your model to be inconsistent. You just feed it your documents in whatever order you want and expect it to work. But by understanding the underlying mechanism of these black-box models better, we can improve them by addressing these limitations,” says Xinyi Wu.
The researchers created a theoretical framework to study how information flows through the machine-learning architecture that forms the backbone of LLMs. They found that certain design choices which control how the model processes input data can cause position bias.
Their experiments revealed that model architectures, particularly those affecting how information is spread across input words within the model, can give rise to or intensify position bias, and that training data also contribute to the problem.
The MIT team’s framework utilizes graph-based analysis to trace dependencies within the attention mechanism, the core component that allows LLMs to understand context. This approach allowed them to identify how attention masking techniques, used to improve computational efficiency, can inadvertently introduce bias towards the beginning of an input sequence.
The study further demonstrates a “lost-in-the-middle” phenomenon, where models perform best when the relevant information is at the beginning or end of a sequence, with accuracy declining as the information moves towards the middle. This has significant implications for applications like legal document review, medical diagnosis, and code assistance, where critical details might be overlooked.
The implications of this research are far-reaching. By understanding the roots of position bias, developers can create more reliable AI systems that are less prone to error and inconsistency. This could lead to improvements in various applications, from chatbots that maintain context during long conversations to medical AI systems that fairly analyze patient data and code assistants that pay attention to the whole program.
“By doing a combination of theory and experiments, we were able to look at the consequences of model design choices that weren’t clear at the time. If you want to use a model in high-stakes applications, you must know when it will work, when it won’t, and why,” says Ali Jadbabaie, professor at MIT.



