Methods and Applications of Graph Neural Networks for Fake News Detection Using AI-Inspired Algorithms

Methods and Applications of Graph Neural Networks for Fake News Detection Using AI-Inspired Algorithms

Arpit Jain (Koneru Lakshmaiah Education Foundation, India), Ishta Rani (Chandigarh University, India), Tarun Singhal (Chandigarh Engineering College, India), Parveen Kumar (Chandigarh University, India), Vinay Bhatia (Chandigarh Engineering College, India), and Ankur Singhal (Chandigarh Engineering College, India)
Copyright: © 2023 |Pages: 16
DOI: 10.4018/978-1-6684-6903-3.ch012
OnDemand:
(Individual Chapters)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Graph data, which often includes a richness of relational information, are used in a vast variety of instructional puzzles these days. Modelling physics systems, detecting fake news on social media, gaining an understanding of molecular fingerprints, predicting protein interfaces, and categorising illnesses all need graph input models. Reasoning on extracted structures, such as phrase dependency trees and picture scene graphs, is essential research that is necessary for other domains, such as learning from non-structural data such as texts and photos. These types of structures include phrase dependency trees and image scene graphs. Graph reasoning models are used for this kind of investigation. GNNs have the ability to express the dependence of a graph via the use of message forwarding between graph nodes. Graph convolutional networks (GCN), graph attention networks (GAT), and graph recurrent networks (GRN) have all shown improved performance in response to a range of deep learning challenges over the course of the last few years.
Chapter Preview
Top

Introduction

Another possible explanation is graph representation learning, which involves the process of understanding how to represent graph nodes, edges, and subgraphs via the use of low-dimensional vectors (Goyal & Ferrara, 2018; Cui et al., 2018a; Hamilton et al., 2017; Zhang et al., 2018a; Cai et al., 2018; Goyal & Ferrara, 2018). Because they depend on hand-engineered features, conventional graph analysis and machine learning approaches are inflexible, time-consuming, and expensive. The SkipGram model is applied to the random walks generated by DeepWalk (Perozzi et al., 2014), which is the first way to graph embedding based on representation learning. DeepWalk was developed by Perozzi and his colleagues. Additionally, substantial progress was achieved in Node2vec, LINE, and TADW. According to Hamilton et al. (2017b), these techniques have two fundamental limitations. First, the encoder does not share any parameters with its offspring nodes. Consequently, the total number of parameters rises proportionately to the total number of nodes, making computation inefficient. Second, direct embedding methods cannot be generalised and cannot handle dynamic graphs.

When creating graph neural networks (GNNs), also known as graph structure data gatherers, CNNs and graph embedding are part of the process. Because of this, they can simulate input and output behaviours that are element-dependent.

The effectiveness of graph neural networks may be evaluated in various ways. In the essay that they published in 2017, Bronstein and his colleagues discuss the problems, prospective solutions, applications, and the future of deep geometric learning. Zhang et al. (2019a) provide a further in-depth analysis and discussion of graph convolutional networks. They explore graph convolution operators, whereas we concentrate on skip connections and pooling operators in GNNs.

The research publications on GNN models carried out by Zhang et al. (2018b), and Chami et al. (2020) are the most current survey studies to be published. Under the findings of Chami et al. (2020), GNNs may be classified as recurrent, convolutional, graph autoencoders or spatial-temporal networks. While Zhang et al. (2018b) provide a comprehensive analysis of graph deep learning approaches, Chami et al. (2020) provide a Graph Encoder-Decoder Model to blend network embedding and graph neural network models. This model was developed in order to improve the accuracy of graph deep learning. This research may be accessed on the web pages maintained by their authors. Our article precisely categorises them and focuses mainly on the more conventional GNN models. In addition, we cover variants of GNN that may be applied to various graphs and their applications in a wide range of business sectors.

Key Terms in this Chapter

GNN: Graph Neural Network (GNN) comes under the family of Neural Networks which operates on the Graph structure and makes the complex graph data easy to understand.

GCN: A GCN is a variant of a convolutional neural network that takes two inputs.

Complete Chapter List

Search this Book:
Reset