Hongbo Bo
What will make misinformation spread: An XAI perspective
Bo, Hongbo; Wu, Yiwen; You, Zinuo; McConville, Ryan; Hong, Jun; Liu, Weiru
Authors
Yiwen Wu
Zinuo You
Ryan McConville
Jun Hong Jun.Hong@uwe.ac.uk
Professor in Artificial Intelligence
Weiru Liu
Contributors
L Longo
Editor
Abstract
Explainable Artificial Intelligence (XAI) techniques can provide explanations of how AI systems or models make decisions, or what factors AI considers when making the decisions. Online social networks have a problem with misinformation which is known to have negative effects. In this paper, we propose to utilize XAI techniques to study what factors lead to misinformation spreading by explaining a trained graph neural network that predicts misinformation spread. However, it is difficult to achieve this with the existing XAI methods for homogeneous social networks, since the spread of misinformation is often associated with heterogeneous social networks which contain different types of nodes and relationships. This paper presents, MisInfoExplainer, an XAI pipeline for explaining the factors contributing to misinformation spread in heterogeneous social networks. Firstly, a prediction module is proposed for predicting misinformation spread by leveraging GraphSAGE with heterogeneous graph convolution. Secondly, we propose an explanation module that uses gradient-based and perturbation-based methods, to identify what makes misinformation spread by explaining the trained prediction module. Experimentally we demonstrate the superiority of MisinfoExplainer in predicting misinformation spread, and also reveal the key factors that make misinformation spread by generating a global explanation for the prediction module. Finally, we conclude that the perturbation-based approach is superior to the gradient-based approach, both in terms of qualitative analysis and quantitative measurements.
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | The World Conference on eXplainable Artificial Intelligence (xAI 2023) |
Start Date | Jul 26, 2023 |
End Date | Jul 28, 2023 |
Acceptance Date | May 25, 2023 |
Online Publication Date | Oct 21, 2023 |
Publication Date | Oct 21, 2023 |
Deposit Date | Jun 27, 2023 |
Publicly Available Date | Oct 22, 2025 |
Publisher | Springer Verlag (Germany) |
Volume | 1902 CCIS |
Pages | 321-337 |
Series Title | Communications in Computer and Information Science |
Book Title | Explainable Artificial Intelligence |
ISBN | 9783031440663 |
DOI | https://doi.org/10.1007/978-3-031-44067-0_17 |
Public URL | https://uwe-repository.worktribe.com/output/10889843 |
Files
This file is under embargo until Oct 22, 2025 due to copyright reasons.
Contact Jun.Hong@uwe.ac.uk to request a copy for personal use.
You might also like
A survey of location inference techniques on Twitter
(2015)
Journal Article
Privacy preserving record linkage in the presence of missing values
(2017)
Journal Article
A novel ensemble learning approach to unsupervised record linkage
(2017)
Journal Article
A collaborative multiagent framework based on online risk-aware planning and decision-making
(2017)
Journal Article
Downloadable Citations
About UWE Bristol Research Repository
Administrator e-mail: repository@uwe.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search