Kuznetsova, Elizaveta; Makhortykh, Mykola; Vziatysheva, Victoria; Stolze, Martha; Baghumyan, Ani; Urman, Aleksandra (20 December 2023). In Generative AI we Trust: Can Chatbots Effectively Verify Political Information? (arXiv). Cornell University 10.48550/arXiv.2312.13096
|
Text
2312.13096.pdf - Published Version Available under License Creative Commons: Attribution (CC-BY). Download (2MB) | Preview |
This article presents a comparative analysis of the ability of two large language model (LLM)-based chatbots, ChatGPT and Bing Chat, recently rebranded to Microsoft Copilot, to detect veracity of political information. We use AI auditing methodology to investigate how chatbots evaluate true, false, and borderline statements on five topics: COVID-19, Russian aggression against Ukraine, the Holocaust, climate change, and LGBTQ+ related debates. We compare how the chatbots perform in high- and low-resource languages by using prompts in English, Russian, and Ukrainian. Furthermore, we explore the ability of chatbots to evaluate statements according to political communication concepts of disinformation, misinformation, and conspiracy theory, using definition-oriented prompts. We also systematically test how such evaluations are influenced by source bias which we model by attributing specific claims to various political and social actors. The results show high performance of ChatGPT for the baseline veracity evaluation task, with 72 percent of the cases evaluated correctly on average across languages without pre-training. Bing Chat performed worse with a 67 percent accuracy. We observe significant disparities in how chatbots evaluate prompts in high- and low-resource languages and how they adapt their evaluations to political communication concepts with ChatGPT providing more nuanced outputs than Bing Chat. Finally, we find that for some veracity detection-related tasks, the performance of chatbots varied depending on the topic of the statement or the source to which it is attributed. These findings highlight the potential of LLM-based chatbots in tackling different forms of false information in online environments, but also points to the substantial variation in terms of how such potential is realized due to specific factors, such as language of the prompt or the topic.
Item Type: |
Working Paper |
---|---|
Division/Institute: |
03 Faculty of Business, Economics and Social Sciences > Social Sciences > Institute of Communication and Media Studies (ICMB) |
UniBE Contributor: |
Makhortykh, Mykola, Vziatysheva, Victoria, Baghumyan, Ani, Urman, Aleksandra |
Subjects: |
000 Computer science, knowledge & systems 300 Social sciences, sociology & anthropology 300 Social sciences, sociology & anthropology > 320 Political science |
Series: |
arXiv |
Publisher: |
Cornell University |
Language: |
English |
Submitter: |
Mykola Makhortykh |
Date Deposited: |
12 Feb 2024 10:12 |
Last Modified: |
12 Feb 2024 10:21 |
Publisher DOI: |
10.48550/arXiv.2312.13096 |
ArXiv ID: |
2312.13096 |
Uncontrolled Keywords: |
artificial intelligence, chatbot, audit, chatGPT, Bard, Bing AI, disinformation, conspiracy theory, misinformation, climate change, Holocaust, war in Ukraine, LGBTQ+, gender, COVID |
BORIS DOI: |
10.48350/192758 |
URI: |
https://boris.unibe.ch/id/eprint/192758 |