Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

G-CutMix: A CutMix-based graph data augmentation method for bot detection in social networks

  • Yan Li ,

    Contributed equally to this work with: Yan Li, Shuhao Shi

    Roles Conceptualization, Writing – original draft

    Affiliation WuXi University, Wuxi, Jiangsu, China

  • Shuhao Shi ,

    Contributed equally to this work with: Yan Li, Shuhao Shi

    Roles Investigation, Methodology

    Affiliation Zhengzhou Information Science and Technology Institue, Zhengzhou, Henan, China

  • Xiaofeng Guo ,

    Roles Formal analysis, Funding acquisition, Investigation

    yanli@cwxu.edu.cn

    Affiliation Zhengzhou Information Science and Technology Institue, Zhengzhou, Henan, China

  • Chunhua Zhou,

    Roles Project administration, Resources

    Affiliation Zhengzhou Information Science and Technology Institue, Zhengzhou, Henan, China

  • Qian Hu

    Roles Software, Validation

    Affiliation Zhengzhou Information Science and Technology Institue, Zhengzhou, Henan, China

Abstract

The CutMix technique is a sophisticated approach for augmenting data in order to train neural network-based image classifiers. Essentially, it involves cutting out a portion of a random image and pasting it into the same location as another image. However, because of the irregularity of graph data, CutMix cannot be directly applied to graph learning. Our paper introduces G-CutMix, a CutMix-based data augmentation approach that we designed specifically for bot detection in social media networks. G-CutMix involves conducting CutMix operations between the original graph and a shuffled graph, which precedes the graph convolution process. The outputs of the graph convolution are then strategically merged with the user representations from both the original and shuffled graphs. Our proposed G-CutMix not only leverages the power of graph convolutions but also introduces a layer of complexity that mimics real-world scenarios where bot behavior can be subtle and varied, making G-CutMix a formidable tool in the arsenal against bot detection. Our experiments confirm that our approach can consistently enhance the performance of bot detection across various GNN architectures, including Graph Convolutional Networks, GraphSAGE, and Graph Attention Networks.

Introduction

Graph neural networks (GNNs) have attained remarkable accuracy in detecting social network bots due to their ability to model interactions between accounts. However, obtaining a significant number of high-quality account annotations in practical scenarios is challenging, and the number of samples available for model training may be limited. Consequently, a well-trained GNN might end up learning random errors or noise rather than the true data distribution [1], which is suboptimal.

Social bot detection techniques are currently classified into three categories: feature-based methods, graph-based methods, and deep learning methods. Feature-based approaches have been in use since the earliest efforts in social bots detection. This approach involves extracting and designing features [2] from users’ metadata, such as Twitter content [3], user data [4], features [5], timeliness, language, and emotions extracted from friends and neighbors’ information. Traditional classifiers are then employed for robot detection. However, social bots can modify their registration information based on the features designed for detection, thereby avoiding feature-based detection methods [6,7].

The graph-based approach utilizes the graph structure of the social network to anticipate the labels of each remaining node. These techniques do not require a significant amount of text and media data that are highly dependent on language. Based on the core algorithms, they can be broadly categorized into two types: the random walk method [810] and the Loopy Belief Propagation (LBP) method [11,12]. Both approaches start with a set of labeled nodes and use semi-supervised learning to predict the labels of unknown nodes. The effective utilization of internal data on social platforms enhances its detection capability significantly higher than the aforementioned method.

The deep learning-based detection method primarily relies on graph representation learning or GNN model, which simultaneously utilizes the user’s attribute and structure information to enhance the detection performance. TrustGCN [13] employs the "friend request-reply" relationship to form a social graph and combines the defensive concepts based on the social graph with GNN to improve the robustness of adversarial attacks. Bot2vec [14] is an improved social robot detection algorithm that is based on Node2vec. It combines a clustering technique with a graph representation learning algorithm for social bot detection. The proposed BotRGCN [15] uses graph convolutional networks to analyze Twitter social graphs, leveraging user attributes and interactions for detecting social bots. SATAR [16] is a self-supervised representation method that leverages user tweets, metadata, and connections to analyze Twitter users.

Recently, data augmentation has proved to be highly effective in convolutional neural networks (CNNs). However, data augmentation in graph neural networks (GNNs) is still sparsely studied. For node classification tasks, Rong et al. [17] proposed a data augmentation technique named DropEdge, which basically involves randomly removing a certain number of edges from the input graph in each training stage. Wang et al. [18] developed MixupForGraph, which starts with standard feed-forward operations and then applies Mixup graph convolutions to nodes in the second stage. Han et al. [19] introduced G-Mixup, a method that interpolates between graphons from different graph classes to enhance graph classification.

We present a novel approach for augmenting data in bot detection using graph learning. Initially, we shuffle the user relationship graph to obtain its isomorphic graph. Next, we cutmix the node features and labels of the two graphs in order. The resultant mixed features are then fed into the graph convolution. The local aggregation-based features produced by the graph convolution are added to the original features of the two views. The outcome is a new layer of features for the two views.

Our cutmix methods can be integrated into well-known GNNs. We assess the efficiency of our approach on real-world bot detection datasets, including Cresci-15 [20], Twibot-20 [21], and MGTAB [22]. In summary, our contribution is threefold:

  • We introduce a novel and versatile data augmentation technique for graph learning based on CutMix. Compared with previous data enhancement methods, our method is more effective.
  • We have extended our method to heterogeneous graphs and demonstrated that integrating information from various relationships between users can significantly enhance the performance of bot detection graph representation learning.
  • Our approach has undergone extensive experimentation on real-world bot detection datasets, which has shown its high efficacy. Compared to earlier data augmentation methods, our approach shows a substantial enhancement in both accuracy and F1-score for bot detection.

Related work

Graph neural network

GNNs process non-Euclidean graph data by generating node-level feature representations through message passing. At layer l + 1, node features are obtained by aggregating information from their 1-hop neighbors at layer l:

Where is the adjacency matrix, and represents the learnable parameter matrix, and σ represents the activation function. AGGREGATE represents the aggregation function that aggregates the hidden representations of neighboring nodes.

GNN-based bot detection

Graph-based methods have proven highly effective in social bot detection [21,22]. START [23] develops a method for learning Twitter user representations, which is then used to refine bot detection. GNNs have recently made significant advances in this area, modeling accounts as nodes and relationships such as friends and followers as edges. Alhosseini et al. [24] were pioneers in applying Graph Convolutional Networks to social bot detection, utilizing account interactions effectively. More recently, NDE-GNN [25] achieves social bot detection by integrating hypergraph-based higher-order neighborhood representations with differential feature enhancement techniques. Shi et al. [26] introduced RF-GNN, which enhances detection accuracy through ensemble learning. However, RF-GNN’s use of averaging to combine base classifiers can reduce effectiveness by not fully capturing the performance variations among different classifiers. Li et al. proposed BotCL [27], a graph contrastive learning-based social bot detection model leveraging multi-view data augmentation to integrate semantic, attribute, and structural features, yet it faces hyperparameter sensitivity and incomplete utilization of heterogeneous social relationships.

Data augmentation

Data augmentation is essential for enhancing model performance, as it modifies the input data to create a more robust learning environment. For instance, in image classification, data augmentation techniques like horizontal flipping and random erasing have been demonstrated to improve performance. Mixup [28] is a powerful data augmentation technique for image classification, involving training a neural network with convex combinations of pairs of images and their labels. Recently, several mixup-based methods have been proposed for graph data to enhance the performance of models. G-Mixup [29] is a data augmentation technique for graph classification that generates a new graph by using a linear combination of two graphs. MixupForGraph [18] combines two graphs in a manner that preserves the graph structure. GraphMix [36] integrates a graph neural network with traditional neural network architectures, using advanced techniques to improve performance.

CutMix [30] represents a unique data augmentation approach that randomly mixes the input data and labels of two samples, training models on the combined data to encourage learning from diverse parts of multiple examples—a strategy more robust than traditional Mixup. Building on this, we developed G-CutMix, a CutMix-based graph data augmentation method designed to enhance GNN performance. Unlike prior Mixup-based graph augmentation methods (e.g., MixupForGraph’s global structure blending), G-CutMix employs local subgraph replacement rather than global interpolation, striking a balance between data diversity and the preservation of local graph structures and topological relationships. This mechanism gives G-CutMix distinct advantages in tasks relying on local information, such as node classification or molecular property prediction. In contrast, MixupForGraph’s global mixing nature may be more suitable for tasks insensitive to fine-grained local structures. With its physical interpretability and ability to retain local structures, G-CutMix emerges as a superior choice for graph data augmentation—addressing the unique challenges of structural data while enhancing model generalization.

Proposed method

Background and motivation

Data augmentation is a straightforward yet effective method to enhance neural network training by expanding the diversity of the training dataset. Mixup-based graph enhancement methods have been proposed, significantly improving the ability of graph representation learning. CutMix and its variants use region-level cut and paste hybrid techniques to force the neural network to pay more attention to the global context of the image rather than just local information, thereby maintaining continuity of information compared to mixup. Unlike MixupForGraph, which linearly interpolates node features, G-CutMix employs region-level feature swapping using a binary mask M (Eq 1). This ensures that local structural patterns (e.g., community-specific interactions) are preserved during augmentation, whereas linear interpolation may blur such patterns. For example, in social graphs, bots often exhibit localized behavioral anomalies (e.g., sudden spikes in follower requests). By retaining intact feature regions, G-CutMix helps GNNs capture these subtle signals more effectively.

In our work, we perform bot detection based on node classification and propose a graph augmentation method named G-CutMix, utilizing a CutMix module to improve the performance of bot detection. The overall framework of G-CutMix is depicted in Fig 1 and comprises of three modules: Graph Shuffle Module, Node CutMix Module, and Attribute Connection Module.

thumbnail
Fig 1. The procedure for training with G-CutMix.

and is isomorphic graph.

https://doi.org/10.1371/journal.pone.0331978.g001

Graph shuffle module

The objective of node shuffling is to disturb the arrangement of nodes and their associated edges within the graph, resulting in the isomorphic graph , which is based on the original graph . The definition of an isomorphic graph can be found in Definition 1. To accomplish this, the node ID in must be shuffled to obtain the shuffled ID . The nodes within should be rearranged according to the order of . The ratio of node shuffling can be denoted by the variable λ, .

Definition 1. Isomorphic graph: Given two graphs and , if there exists a bijection m: such that for all is equivalent to , then and are isomorphic.

Node CutMix module

Sample pairs are created by selecting corresponding numbered nodes from and . For instance, the sample pair constructed from the i-th node in both graphs is represented as , where N denotes the number of nodes within the graph. The Node CutMix Module is depicted in Fig 2.

CutMix the representations of sample pairs according to Eq 1:

(1)

here represents a binary mask of length H, where each element indicates whether to drop out or retain the corresponding entry from two vectors. is the hyperparameter that controls the ratio of 1 in mask. and and pass through the Graph CutMix Module to get .

Attribute connection module

The Attribute Connection Module comprises a full connection layer and is utilized to preserve the original attributes of the graph. The output obtained by via the graph convolution layer (GConv) and the initial representations are obtained by adding features and through the linear layer, resulting in obtaining and :

(2)(3)

Afterward, the resulting and are fed into the Node CutMix Module in the subsequent layer. Once CutMix is employed to enhance the training distribution features, corresponding labels should be mixed while computing the loss function:

(4)

Experiment

Dataset

We assess the performance of Twitter bot detection models on three datasets that possess graph structures:

  • Cresci-15 [20] is a dataset with 5,301 users labeled as either genuine or automated accounts, and it includes details on their follower and friend relationships.
  • Twibot-20 [21] is a dataset comprising 229,580 users and 227,979 edges, with 11,826 accounts labeled as either genuine or automated. It includes detailed information on follower and friend relationships among these users.
  • MGTAB [22] is a comprehensive dataset designed for machine account detection, featuring over 1.5 million users and 130 million tweets. It includes data on seven types of relationships between these users and labels 10,199 accounts as either genuine or bots. The dataset details seven types of relationships among users and labels 10,199 accounts as either genuine or bots.

We build user social graphs using labeled users from each dataset. For MGTAB, we incorporate 20 high-information-gain user attributes and 768-dimensional tweet features from BERT. In Twibot-20, we use 17 user attributes, 768-dimensional description features from BERT, and tweet features. For Cresci-15, we use 6 user attributes, 768-dimensional description features from BERT, and tweet features. The details of these datasets are presented in Table 1. For all datasets, we partition the data randomly into training, validation, and testing sets in a 0.5:1:8.5 ratio.

thumbnail
Table 1. Statistics of bot detection datasets used in the paper.

https://doi.org/10.1371/journal.pone.0331978.t001

Baseline methods

To evaluate the effectiveness of our proposed method, we compare it against commonly used GNN methods, including:

GCN: [31] is a method used for semi-supervised learning on graph-structured data, focusing on localized graph convolution.

SAGE: [32] comprises sampling and aggregation. Initially, it selects neighboring nodes based on their connections and then merges information from these neighbors through a series of multi-layer aggregation functions. The fused information is used to predict the node label.

GAT: [33] utilizes an attention mechanism to assign weights to node neighborhoods. This adaptive weighting of neighbors enhances the performance of graph neural networks.

JK-Nets: [34] mitigates over-smoothing concerns in deep GNNs by harnessing varying neighborhood scopes for flexible aggregation.

LA-GCN: [35] introduces a layer-wise attention mechanism, enabling the network to discern the significance of each layer’s output in the ultimate node representation.

RF-GNN: [26] is a method that randomly deletes a portion of edges from the input graph at each training stage.

DropEdge: [17] is a technique that randomly removes a certain number of edges from the input graph during each training stage.

MixupForGraph: [18] is a two-stage method for GNNs. It first performs standard feed-forward operations, then incorporates Mixup-based techniques to blend information from different nodes during each layer of the network.

Configurations

All of our models are constructed with two layers of graph convolution layers with ReLU activation and a dropout rate of 0.5. We employ the Cross-Entropy loss function and the Adam optimizer for model optimization, setting the learning rate to 1e-3 and a weight decay to 5e-4. For the Cresci-15 [20], Twibot-20 [21], and MGTAB [22] datasets, we utilize followers and friends user relationships, respectively. We follow [21] to extract and construct user feature attributes. We set α (the hyperparameter for the adaptive threshold) to 0.4 and train the model for 200 epochs. Attention heads in GAT and RGAT are configured to 4.

Main results

In this section, we investigate the performance of G-CutMix using follower and friend relationships in bot detection tasks, respectively. In our experiments, we created social relationship graphs using followers and friends relationships and then utilized graph neural network models to detect machine accounts. Each baseline was executed five times with varying initializations to mitigate randomness. The experimental outcomes are displayed in Tables 2 and 3.

thumbnail
Table 2. Performance of bot detection using followers relationship graph. Acc and F1 stands for accuracy and F1-score (macro), respectively. Boldface letters are used to mark the best results.

https://doi.org/10.1371/journal.pone.0331978.t002

thumbnail
Table 3. Performance of bot detection using friends relationship graph. Acc and F1 stands for accuracy and F1-score (macro), respectively. Boldface letters are used to mark the best results.

https://doi.org/10.1371/journal.pone.0331978.t003

Utilizing both followers relationship graphs and friends relationship graphs, our proposed G-CutMix for bots detection consistently outperforms baseline methods. Employing GCN, SAGE, and GAT as backbone models, and comparing with graph data augmentation techniques like DropEdge and MixupForGraph, our G-CutMix, through multi-level data enhancement, achieves superior performance.

Randomly selecting 1,000 nodes, the t-SNE dimensionality reduction and visualization of the hidden layer features obtained by various methods are depicted in Fig 3, with labels 1 and 0 representing bots and human accounts, respectively. The embeddings obtained by the GCN model exhibit the highest degree of overlap in the feature space, as evidenced by the fact that most green points are occluded by orange points. DropEdge and MixupGCN demonstrate better distinguishability in the feature space compared to GCN. The embeddings generated by CutMix show the lowest overlap in the feature space and achieve the highest distinguishability. In the case of the GCN model, the model’s representational capacity is insufficient, resulting in a significant overlap of the hidden layer features between bots and human accounts, which indicates poor separability. DropEdge randomly discards edges, but the effect on feature enhancement is not pronounced. In contrast, G-CutMix and MixupForGraph enhance the separability of features through data augmentation of node features. G-CutMix outperforms both DropEdge and MixupForGraph in terms of effectiveness.

thumbnail
Fig 3. The t-SNE visualizations of the output results from various methods.

(a) GCN. (b) DropEdge. (c) MixupForGrap. (d) CutMix.

https://doi.org/10.1371/journal.pone.0331978.g003

Discussion

Training set sizing

In the field of bots detection, obtaining a large-scale labeled dataset is often challenging. In this section, we showcase that our G-CutMix approach results in more significant improvements when trained on smaller datasets. The validation set and test set are the same as CutMix for Heterogeneous Graph section, changing the training set size from 1% to 5%. The results are reported in Fig 4.

thumbnail
Fig 4. Accuracy of different training sets using different relationship graph.

(a) Results based on friends relationship. (b) Results based on followers relationship.

https://doi.org/10.1371/journal.pone.0331978.g004

Upon examining the performance of G-CutMix and the original GNN, it is evident that our G-CutMix approach enhances the performance of GCN and GAT across various training set sizes. Our method shows more significant improvements for smaller training sets because it is challenging for the model to achieve sufficient training with limited training data.

Parameters sensitivity analysis

The parameter α controls the CutMix ratio, and in this section, we analyze the effect of α in our method. When α equals 0, the node features of the two graph views are not augmented with CutMix. When a equals 1, the mix ratio of features in and is maximized, with features directly exchanged.

Experiments in Fig 5 demonstrate that achieves optimal performance across all datasets (MGTAB, Cresci-15, Twibot-20). Specifically, varying α from 0.1 to 0.9 on MGTAB results in accuracy fluctuations within , suggesting parameter robustness. We further propose empirical guidelines: datasets with sparse graphs (e.g., Twibot-20) achieve best performance with , whereas is preferred for dense graphs like MGTAB.

CutMix for heterogeneous graph

In the field of social bots detection, it has been observed that using multiple relations instead of just a single relation leads to better performance. Our proposed approach can be extended to multi-relation graphs, where multiple relations are used simultaneously. To this end, we constructed a social network graph with two relations, namely followers and friends, using a machine detection dataset. The comparison results of G-CutMix with classic heterogeneous graph neural networks RGCN [37] and RGAT [29] are presented in Table 4. This approach proves to be advantageous since G-CutMix enables the model to generalize better to unseen data by leveraging information from multiple relations.

thumbnail
Table 4. Performance on multi-relation graphs using both friendship and follower relations. Boldface letters indicate the best results.

https://doi.org/10.1371/journal.pone.0331978.t004

The consistent superiority of G-CutMix-enhanced models, average 1.91% F1 improvements on MGTAB, 1.95% on Cresci-15, 3.82% on Twibot-20 stems from its ability to synergize follower and friend relations. Unlike baseline RGCN/RGAT that process relations sequentially, our method’s isomorphic shuffling and feature fusion create implicit cross-relational attention—for instance, amplifying signals where follower-friend reciprocity indicates coordinated bot behavior. While both RGCN and RGAT benefit from G-CutMix, the greater improvements with RGAT highlight our method’s compatibility with attention mechanisms. The learnable merging weights in G-CutMix likely synergize with RGAT’s edge-specific attention, enabling adaptive reweighting of mixed features.

Ablation study

We divided the dataset into training, validation, and test sets as outlined in Main Results. To evaluate the impact of each module in G-CutMix, we used the followers relationship graph. Specifically, we conducted experiments by removing the Node Shuffle Module, Graph CutMix Module, and Attribute Connection Module individually. The settings “w/o shuffle", “w/o CutMix", and “w/o attribute" correspond to our original method without the Node Shuffle Module, Graph CutMix Module, and Attribute Connection Module, respectively.

To investigate the effects of node shuffle and CutMix in our method, we removed the Node Shuffle Module and set the mixing parameter in CutMix to be 0. The results of these experiments are presented in Table 5.

thumbnail
Table 5. The accuracy of G-CutMix and its variants. Boldface letters are used to mark the best results.

https://doi.org/10.1371/journal.pone.0331978.t005

The ablation study results in Table 5 critical interdependencies between G-CutMix’s components and dataset characteristics. The most pronounced performance degradation occurs when removing the Attribute Connection Module (average 3.8% F1 drop across datasets), particularly severe in Twibot-20 (6.2% accuracy decline for GCN), suggesting that social bots’ attribute camouflage strategies - such as profile metadata manipulation - require explicit attribute correlation modeling to detect. Interestingly, while Node Shuffle removal impacts MGTAB most significantly (1.5-2.5% accuracy reduction), its effect diminishes in Twibot-20 where temporal behavioral patterns dominate, implying that structural isomorphism preservation becomes less critical when bots exhibit strong activity sequence signatures. These findings collectively demonstrate that G-CutMix’s power emerges from the synergistic combination of its components rather than any single module.

Computational complexity

On MGTAB with a GCN backbone, G-CutMix increases training time by 20% compared to vanilla GCN, as shown in Table 6. These results confirm that G-CutMix’s performance gains outweigh its modest computational cost, making it suitable for real-world deployment.

thumbnail
Table 6. Computational complexity comparisons for different methods.

https://doi.org/10.1371/journal.pone.0331978.t006

Conclusions

In this paper, we introduce a novel graph data augmentation method based on CutMix to enhance the performance of bots detection in social networks. Our approach involves mixing the node features and labels of two graphs in sequence using CutMix, followed by feeding the mixed features into a graph convolution. The convolution generates local aggregation-based features, which are then combined with the original features of the two graphs to produce a new layer of features. Through extensive experiments, we demonstrate the effectiveness of our proposed approach, G-CutMix, in detecting bots on social networks, especially in scenarios with limited labeled nodes.

References

  1. 1. Zhang C, Bengio S, Hardt M, Recht BH, Vinyals O. Understanding deep learning (still) requires rethinking generalization. Commun ACM. 2021;64(3):107–15.
  2. 2. Davis CA, Varol O, Ferrara E, Flammini A, Menczer F. BotOrNot: A system to evaluate social bots. In: Proceedings of the 25th international conference on world wide web, Montreal, Canada; 2016. p. 273–4.
  3. 3. Miller Z, Dickinson B, Deitrick W, Hu W, Wang AH. Twitter spammer detection using data stream clustering. Inform Sci. 2014;260:64–73.
  4. 4. D’Andrea E, Ducange P, Lazzerini B, Marcelloni F. Real-time detection of traffic from Twitter stream analysis. IEEE Trans Intell Transp Syst. 2015;16(4):2269–83.
  5. 5. Yang C, Harkreader R, Gu G. Empirical evaluation and new design for fighting evolving twitter spammers. IEEE TransInformForensic Secur. 2013;8(8):1280–93.
  6. 6. Cresci S, Di Pietro R, Petrocchi M, Spognardi A, Tesconi M. The paradigm-shift of social spambots: Evidence, theories, and tools for the arms race. In: Companion proceedings of the 26th international conference on world wide web, Perth, Australia; 2017. p. 963–72.
  7. 7. Feng S, Tan Z, Wan H, Wang N, Chen Z, Zhang B, et al. TwiBot-22: Towards graph-based Twitter bot detection. arXiv preprint; 2022. https://doi.org/10.48550/arXiv.2206.04564
  8. 8. Yu H, Kaminsky M, Gibbons PB, Flaxman AD. SybilGuard: Defending against sybil attacks via social networks. IEEE/ACM Trans Netw. 2008;16(3):576–89.
  9. 9. Yu H, Gibbons PB, Kaminsky M, Xiao F. SybilLimit: A near-optimal social network defense against sybil attacks. In: 2008 IEEE symposium on security and privacy (sp 2008); 2008. p. 3–17. https://doi.org/10.1109/sp.2008.13
  10. 10. Liu Y, Ji S, Mittal P. SmartWalk: Enhancing social network security via adaptive random walks. In: Proceedings of the 2016 ACM SIGSAC conference on computer and communications security; 2016. p. 134–51.
  11. 11. Danezis G, Mittal P. SybilInfer: Detecting Sybil nodes using social networks. In: Proceedings of the network and distributed system security symposium (NDSS), San Diego, California, USA; 2009. p. 324–42.
  12. 12. Wang B, Le Z, Gong NZ. SybilSCAR: Sybil detection in online social networks via local rule based propagation. In: 2017 IEEE conference on computer communications (INFOCOM); 2017. p. 421–40.
  13. 13. Sun Y, Yang Z, Dai Y. TrustGCN: Enabling graph convolutional network for robust Sybil detection in OSNs. In: 2020 IEEE/ACM international conference on advances in social networks analysis and mining (ASONAM); 2020. p. 567–91.
  14. 14. Pham P, Nguyen LTT, Vo B, Yun U. Bot2Vec: A general approach of intra-community oriented representation learning for bot detection in different types of social networks. Inform Syst. 2021;101771.
  15. 15. Feng S, Wan H, Wang N, Luo M. BotRGCN: Twitter bot detection with relational graph convolutional networks. In: Proceedings of the 2021 IEEE/ACM international conference on advances in social networks analysis and mining (ASONAM); 2021. p. 236–9.
  16. 16. Feng S, Wan H, Wang N, Li J, Luo M. SATAR: A self-supervised approach to Twitter account representation learning and its application in bot detection. In: Proceedings of the 30th ACM international conference on information and knowledge management (CIKM); 2021. p. 3808–17.
  17. 17. Rong Y, Huang W, Xu T, Huang J. DropEdge: Towards deep graph convolutional networks on node classification. In: International Conference on Learning Representations (ICLR); 2020. p. 512–31.
  18. 18. Wang Y, Wang W, Liang Y, Cai Y, Hooi B. Mixup for node and graph classification. In: Proceedings of the web conference 2021 ; 2021. p. 3220–2248.
  19. 19. Han X, Jiang Z, Liu N, Hu X. G-Mixup: Graph data augmentation for graph classification. In: Proceedings of the 39th international conference on machine learning (ICML); 2022. p. 8230–48.
  20. 20. Cresci S, Di Pietro R, Petrocchi M, Spognardi A, Tesconi M. Fame for sale: Efficient detection of fake Twitter followers. Decision Support Syst. 2015;80:56–71.
  21. 21. Feng S, Wan H, Wang N, Li J, Luo M. TwiBot-20: A comprehensive Twitter bot detection benchmark. In: Proceedings of the 30th ACM international conference on information and knowledge management (CIKM), 2021. 4485–94.
  22. 22. Shi S, Qiao K, Chen J, Yang S, Yang J, Song B, et al. MGTAB: A multi-relational graph-based Twitter account detection benchmark; 2023. https://doi.org/arXiv:2301.01123
  23. 23. Feng S, Wan H, Wang N, Li J, Luo M. SATAR: A self-supervised approach to Twitter account representation learning and its application in bot detection. In: Proceedings of the 30th ACM international conference on information and knowledge management (CIKM); 2021. p. 148–53.
  24. 24. Ali Alhosseini S, Bin Tareaf R, Najafi P, Meinel C. Detect me if you can: Spam bot detection using inductive representation learning. In: Companion proceedings of the 2019 world wide web conference; 2019. p. 148–53. https://doi.org/10.1145/3308560.3316504
  25. 25. Shi S, Li Y, Liu Z, Chen C, Chen J, Yan B. Neighborhood difference-enhanced graph neural network based on hypergraph for social bot detection. In: Proceedings of the Chinese conference on pattern recognition and computer vision (PRCV); 2025.
  26. 26. Shi S, Qiao K, Yang J, Song B, Chen J, Yan B. RF-GNN: Random forest boosted graph neural network for social bot detection; 2023. https://arxiv.org/abs/2304.08239
  27. 27. Li Y, Li Z, Gong D, Hu Q, Lu H. BotCL: A social bot detection model based on graph contrastive learning. Knowl Inform Syst. 2024;66(9).
  28. 28. Zhang H, Cisse M, Dauphin YN, Lopez-Paz D. Mixup: Beyond empirical risk minimization. arXiv preprint; 2017. https://doi.org/10.48550/arXiv.1710.09412
  29. 29. Han X, Jiang Z, Liu N, Hu X. G-mixup: Graph data augmentation for graph classification. In: Proceedings of the 39th international conference on machine learning (ICML); 2022. p. 8230–48.
  30. 30. Yun S, Han D, Oh SJ, Chun S, Choe J, Yoo Y. CutMix: Regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE/CVF international conference on computer vision (ICCV); 2019. p. 6023–32.
  31. 31. Kipf TN, Welling M. Semi-supervised classification with graph convolutional networks. arXiv preprint; 2016. https://doi.org/10.48550/arXiv.1609.02907
  32. 32. Hamilton W, Ying Z, Leskovec J. Inductive representation learning on large graphs. In: Advances in neural information processing systems 30; 2017.
  33. 33. Velikovi P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y. Graph attention networks; 2017. https://arxiv.org/abs/1710.10903
  34. 34. Xu K, Li C, Tian Y, Sonobe T, Kawarabayashi K, Jegelka S. Representation learning on graphs with jumping knowledge networks. In: Proceedings of the 35th international conference on machine learning (ICML); 2018. p. 5453–62.
  35. 35. Liu S, Ying R, Dong H, Li L, Xu T, Rong Y, et al. Local augmentation for graph neural networks. In: Proceedings of the 39th international conference on machine learning (ICML); 2022. p. 14054–72.
  36. 36. Verma V, Qu M, Kawaguchi K, Lamb A, Bengio Y, Kannala J, et al. GraphMix: Improved training of GNNs for semi-supervised learning. In: Proceedings of the 35th AAAI conference on artificial intelligence; 2021. p. 10024–32.
  37. 37. Schlichtkrull M, Kipf TN, Bloem P, Van Den Berg R, Titov I, Welling M. Modeling relational data with graph convolutional networks. In: Proceedings of the 15th extended semantic web conference (ESWC), Heraklion, Crete, Greece; 2018. p. 593–607.