Although Large Language Models (LLMs) are frequently used to generate text, their ability to maintain neutrality when addressing sensitive topics remains a concern. By supplying inputs with predetermined positions and examining the replies produced, this study explores the positions taken by three LLMs - Mixtral-8x7B, Gemma2- 9B, and LLaMA-3.1-8B - on particular issues, including abortion, death penalty, marijuana legalization, nuclear energy, and feminism. The stance of each response was measured, revealing that the models exhibit polarization toward specific positions on these topics. The results point to a serious vulnerability in the models’ ability to remain neutral since their answers frequently reflect a prevailing viewpoint in sensitive contexts. This behavior highlights bias and raises questions about how it can affect users, who might be trapped in a cognitive filter bubble influenced by the model’s polarized responses. This work sheds light on the challenges LLMs’ bias poses, emphasizing the need for strategies to ensure their neutrality and mitigate the risks associated with reinforcing distorted perspectives during user interactions.

Cognitive filter bubble: investigating bias and neutrality vulnerabilities of LLMs in sensitive contexts / Di Gisi, Maria; Fenza, Giuseppe; Gallo, Mariacristina; Loia, Vincenzo; Stanzione, Claudio. - 3962:(2025). ( ITASEC & SERICS 2025 - Joint National Conference on Cybersecurity 2025 Bologna, Italy 03-08/12/2025).

Cognitive filter bubble: investigating bias and neutrality vulnerabilities of LLMs in sensitive contexts

Di Gisi Maria;
2025

Abstract

Although Large Language Models (LLMs) are frequently used to generate text, their ability to maintain neutrality when addressing sensitive topics remains a concern. By supplying inputs with predetermined positions and examining the replies produced, this study explores the positions taken by three LLMs - Mixtral-8x7B, Gemma2- 9B, and LLaMA-3.1-8B - on particular issues, including abortion, death penalty, marijuana legalization, nuclear energy, and feminism. The stance of each response was measured, revealing that the models exhibit polarization toward specific positions on these topics. The results point to a serious vulnerability in the models’ ability to remain neutral since their answers frequently reflect a prevailing viewpoint in sensitive contexts. This behavior highlights bias and raises questions about how it can affect users, who might be trapped in a cognitive filter bubble influenced by the model’s polarized responses. This work sheds light on the challenges LLMs’ bias poses, emphasizing the need for strategies to ensure their neutrality and mitigate the risks associated with reinforcing distorted perspectives during user interactions.
2025
Large Language Models, Vulnerability, Bias, Information disorder
File in questo prodotto:
File Dimensione Formato  
paper17.pdf

accesso aperto

Descrizione: Cognitive Filter Bubble: Investigating Bias and Neutrality Vulnerabilities of LLMs in Sensitive Contexts
Tipologia: Versione Editoriale (PDF)
Licenza: Creative commons
Dimensione 2.21 MB
Formato Adobe PDF
2.21 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.11771/39699
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
social impact