Home
Scholarly Works
Where You Say Matters: A Study of Positional Bias...
Conference

Where You Say Matters: A Study of Positional Bias of Small LLMs

Abstract

Large language models (LLMs) have performed very well across various natural language tasks. These are now increasingly being integrated into handheld and IoT devices. Understanding inherent biases and responses to varying prompting techniques becomes more important. This study examines positional bias using multiple-choice questions and answers using seven small large language models, including models from Microsoft (Phi), Meta (Llama), Google (Gemma), Deepseek and Alibaba Cloud (Qwen). Using the OpenBookQA dataset, our study shows statistically significant results, such as Phi models that demonstrate a preference for D, whereas Deepseek has a significant bias to A. Standard prompting techniques such as Chain of Thought and few-shot prompting do not improve performance and, in some cases, degrade performance. These results highlight the need for bias-aware model selection, especially as these models are increasing in our everyday lives.

Authors

Vella S; Sharieh S; Hussain F; Ferworn A

Volume

00

Pagination

pp. 0640-0647

Publisher

Institute of Electrical and Electronics Engineers (IEEE)

Publication Date

May 30, 2025

DOI

10.1109/aiiot65859.2025.11105359

Name of conference

2025 IEEE World AI IoT Congress (AIIoT)
View published work (Non-McMaster Users)

Contact the Experts team