Large language models (LLMs) are rapidly integrating into numerous facets of modern life, yet concerns regarding inherent biases within these systems persist. Recent research investigates how diverse gender identities perceive and evaluate these biases, alongside assessments of accuracy and trustworthiness, when interacting with LLMs. Aimen GABA, Emily Wall, Tejas Ramkumar Babu, Yuriy Brun, Kyle WM Hall, and Cindy Xiong Bearfield, representing institutions including the University of Massachusetts Amherst, Emory University, Georgia Institute of Technology, and TD Bank, present their findings in “Bias, Accuracy, and Trust: Gender-Diverse Perspectives on Large Language Models”. Through in-depth interviews, the study examines how gendered prompts influence LLM responses and how users from varied gender backgrounds assess the resulting information, revealing nuanced differences in perception and trust.
Large language models (LLMs) are increasingly integrated into daily life, yet concerns regarding inherent biases within these systems necessitate careful scrutiny. Recent research investigates how individuals with diverse gender identities perceive bias, accuracy, and trustworthiness in LLMs, employing qualitative methods. Researchers conducted 25 in-depth interviews with participants identifying as non-binary/transgender, male, and female, assessing responses to both gendered and neutral prompts. Findings demonstrate that gendered prompts frequently elicit identity-specific responses, with non-binary participants reporting a disproportionate experience of condescending or stereotypical portrayals, highlighting a critical need for greater inclusivity in LLM development.
The research reveals a consistent presence of bias within LLMs, as perceived by a diverse range of participants, and this bias demonstrably affects user trust and response framing. Perceived accuracy remains relatively consistent across all gender groups, although participants consistently identify errors in technical subjects and creative tasks, indicating that while bias is a significant concern, the fundamental capability of LLMs to provide factually correct information is not inherently gendered. Trustworthiness, however, exhibits gender-based variation; male participants generally express higher levels of trust, particularly regarding system performance, while non-binary participants demonstrate greater trust based on perceived performance quality. This suggests that trust is not solely determined by the model’s accuracy, but is also heavily influenced by the user’s lived experience and their perception of the model’s inclusivity.
LLMs do not operate neutrally, but actively incorporate and perpetuate existing societal biases, significantly impacting user experience. The study’s findings demonstrate that gendered prompts frequently elicit identity-specific responses, confirming that LLMs are not unbiased arbiters of information. This is particularly pronounced for non-binary participants, who report encountering responses that are either dismissive or reinforce harmful stereotypes.
Future research should focus on developing robust evaluation metrics specifically designed to detect and quantify subtle forms of bias in LLM outputs. Investigating the impact of intersectional identities—considering the combined effects of gender, race, and other social categories—is also crucial, as biases rarely operate in isolation. Exploring how these biases manifest across different cultural contexts and linguistic variations will further refine our understanding of the problem. Researchers must prioritise the creation of datasets that accurately represent the diversity of human experiences, requiring ongoing collaboration with marginalised communities. Developing algorithms that are less susceptible to bias and more transparent in their decision-making processes is also crucial, necessitating a multidisciplinary approach involving computer scientists, social scientists, and ethicists. Furthermore, ongoing monitoring and evaluation of LLMs are essential to identify and address any emerging biases.
This study contributes to the fields of Computer-Supported Cooperative Work (CSCW) and Human-Computer Interaction (HCI) by emphasising the critical importance of incorporating gender-diverse perspectives into the development of LLMs and artificial intelligence more broadly. By acknowledging and addressing the inherent biases within these systems, researchers and developers can work towards creating AI that is not only intelligent but also inclusive, equitable, and trustworthy for all users. This research will enable the development of more equitable and inclusive AI systems, ensuring that these systems benefit all members of society.
👉 More information
🗞 Bias, Accuracy, and Trust: Gender-Diverse Perspectives on Large Language Models
🧠 DOI: https://doi.org/10.48550/arXiv.2506.21898
