Mixture of Message Passing Experts with Routing Entropy Regularization for Node Classification

arXiv — cs.LGThursday, November 13, 2025 at 5:00:00 AM
Graph neural networks (GNNs) have made significant strides in graph-based learning tasks, but they often struggle with heterophilous structures where connected nodes differ in features and labels. To tackle this challenge, researchers introduced GNNMoE, a novel framework that employs a mixture of message-passing experts combined with routing entropy regularization. This approach allows for adaptive representation learning at the node level. GNNMoE separates message passing into propagation and transformation operations, integrating them through expert networks guided by a hybrid routing mechanism. The routing entropy regularization dynamically adjusts soft weighting and top-k routing, enabling the model to flexibly adapt to various neighborhood contexts. Extensive experiments conducted on twelve benchmark datasets demonstrate that GNNMoE consistently outperforms state-of-the-art node classification methods while maintaining scalability and interpretability. This advancement represents …
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
Explicit Multimodal Graph Modeling for Human-Object Interaction Detection
PositiveArtificial Intelligence
Recent advancements in Human-Object Interaction (HOI) detection have seen the rise of Transformer-based methods. However, these methods do not adequately model the relational structures essential for recognizing interactions. This paper introduces Multimodal Graph Network Modeling (MGNM), which utilizes Graph Neural Networks (GNNs) to better capture the relationships between human-object pairs, thereby enhancing HOI detection through a four-stage graph structure and a multi-level feature interaction mechanism.
Multi-View Polymer Representations for the Open Polymer Prediction
PositiveArtificial Intelligence
The article discusses a novel approach to polymer property prediction using a multi-view design that incorporates various representations. The system combines four families of representations: tabular RDKit/Morgan descriptors, graph neural networks, 3D-informed representations, and pretrained SMILES language models. This ensemble method achieved a public mean absolute error (MAE) of 0.057 and a private MAE of 0.082, ranking 9th out of 2241 teams in the Open Polymer Prediction Challenge at NeurIPS 2025.
Enhancing Graph Representations with Neighborhood-Contextualized Message-Passing
PositiveArtificial Intelligence
Graph neural networks (GNNs) are essential for analyzing relational data, categorized into convolutional, attentional, and message-passing variants. The standard message-passing approach, while expressive, overlooks the rich contextual information from the broader local neighborhood, limiting its ability to learn complex relationships. This article introduces a new framework called neighborhood-contextualized message-passing (NCMP) to address this limitation, enhancing the expressivity and efficiency of GNNs.
Posterior Label Smoothing for Node Classification
PositiveArtificial Intelligence
Label smoothing is a regularization technique in machine learning that has not been extensively applied to node classification in graph-structured data. This study introduces posterior label smoothing, a method that generates soft labels based on neighborhood labels. The approach adapts to various graph properties and has been tested on 10 benchmark datasets, showing consistent improvements in classification accuracy and reduced overfitting during training.
Flow-Attentional Graph Neural Networks
PositiveArtificial Intelligence
Graph Neural Networks (GNNs) are crucial for analyzing graph-structured data, but current models overlook the conservation laws relevant to physical resource flows, such as electrical currents in power grids. To improve performance, a new approach called flow attention is introduced, which aligns with Kirchhoff's first law. Experiments on electronic circuits and power grids demonstrate that this method enhances the effectiveness of attention-based GNNs in classification and regression tasks.
Hypergraph Neural Network with State Space Models for Node Classification
PositiveArtificial Intelligence
Recent advancements in graph neural networks (GNNs) have highlighted their effectiveness in node classification tasks. However, traditional GNNs often neglect role-based characteristics that can enhance node representation learning. To overcome these limitations, a new model called the hypergraph neural network with state space model (HGMN) has been proposed, integrating role-aware representations and employing hypergraph construction techniques to capture complex relationships among nodes.
Dynamic Deep Graph Learning for Incomplete Multi-View Clustering with Masked Graph Reconstruction Loss
NeutralArtificial Intelligence
The article presents a novel approach to incomplete multi-view clustering (IMVC) through Dynamic Deep Graph Learning with Masked Graph Reconstruction Loss. It highlights the limitations of existing methods, particularly their reliance on K-Nearest Neighbors (KNN) and Mean Squared Error (MSE) loss, which can introduce noise and reduce graph robustness. The proposed method aims to enhance the effectiveness of IMVC by addressing these challenges, thereby contributing to the advancement of Graph Neural Networks (GNNs) in this field.
Heterogeneous Attributed Graph Learning via Neighborhood-Aware Star Kernels
PositiveArtificial Intelligence
The article presents the Neighborhood-Aware Star Kernel (NASK), a new graph kernel for attributed graph learning. Attributed graphs, which feature irregular topologies and a mix of numerical and categorical attributes, are prevalent in areas like social networks and bioinformatics. NASK utilizes an exponential transformation of the Gower similarity coefficient to efficiently model these attributes and incorporates multi-scale neighborhood structural information through star substructures enhanced by Weisfeiler-Lehman iterations. The theoretical proof confirms that NASK is positive definite.