In Generative AI we Trust: Can Chatbots Effectively Verify Political Information?

Kuznetsova, Elizaveta; Makhortykh, Mykola; Vziatysheva, Victoria; Stolze, Martha; Baghumyan, Ani; Urman, Aleksandra (20 December 2023). In Generative AI we Trust: Can Chatbots Effectively Verify Political Information? (arXiv). Cornell University 10.48550/arXiv.2312.13096

[img]
Preview
Text
2312.13096.pdf - Published Version
Available under License Creative Commons: Attribution (CC-BY).

Download (2MB) | Preview

This article presents a comparative analysis of the ability of two large language model (LLM)-based chatbots, ChatGPT and Bing Chat, recently rebranded to Microsoft Copilot, to detect veracity of political information. We use AI auditing methodology to investigate how chatbots evaluate true, false, and borderline statements on five topics: COVID-19, Russian aggression against Ukraine, the Holocaust, climate change, and LGBTQ+ related debates. We compare how the chatbots perform in high- and low-resource languages by using prompts in English, Russian, and Ukrainian. Furthermore, we explore the ability of chatbots to evaluate statements according to political communication concepts of disinformation, misinformation, and conspiracy theory, using definition-oriented prompts. We also systematically test how such evaluations are influenced by source bias which we model by attributing specific claims to various political and social actors. The results show high performance of ChatGPT for the baseline veracity evaluation task, with 72 percent of the cases evaluated correctly on average across languages without pre-training. Bing Chat performed worse with a 67 percent accuracy. We observe significant disparities in how chatbots evaluate prompts in high- and low-resource languages and how they adapt their evaluations to political communication concepts with ChatGPT providing more nuanced outputs than Bing Chat. Finally, we find that for some veracity detection-related tasks, the performance of chatbots varied depending on the topic of the statement or the source to which it is attributed. These findings highlight the potential of LLM-based chatbots in tackling different forms of false information in online environments, but also points to the substantial variation in terms of how such potential is realized due to specific factors, such as language of the prompt or the topic.

Item Type:

Working Paper

Division/Institute:

03 Faculty of Business, Economics and Social Sciences > Social Sciences > Institute of Communication and Media Studies (ICMB)

UniBE Contributor:

Makhortykh, Mykola, Vziatysheva, Victoria, Baghumyan, Ani, Urman, Aleksandra

Subjects:

000 Computer science, knowledge & systems
300 Social sciences, sociology & anthropology
300 Social sciences, sociology & anthropology > 320 Political science

Series:

arXiv

Publisher:

Cornell University

Language:

English

Submitter:

Mykola Makhortykh

Date Deposited:

12 Feb 2024 10:12

Last Modified:

12 Feb 2024 10:21

Publisher DOI:

10.48550/arXiv.2312.13096

ArXiv ID:

2312.13096

Uncontrolled Keywords:

artificial intelligence, chatbot, audit, chatGPT, Bard, Bing AI, disinformation, conspiracy theory, misinformation, climate change, Holocaust, war in Ukraine, LGBTQ+, gender, COVID

BORIS DOI:

10.48350/192758

URI:

https://boris.unibe.ch/id/eprint/192758

Actions (login required)

Edit item Edit item
Provide Feedback