In this section, we study the integration of edge features to enhance the performance of Graph Convolutional Networks (GCN), Graph Attention Networks (GAT), and GraphSAGE (SAGE) models. The conventional approach without incorporating edge features follows the standard procedure. Initially, employing one of the GNN models mentioned (GCN, GAT, or SAGE), node embeddings denoted as
In contrast, our proposed model, PROXI-GNN, we introduce a novel approach by incorporating the edge features proposed in our study directly into the prediction process. Specifically, within the predictor, our model takes the edge features of edge
In Figure 1, we present a comparative analysis between the standard algorithm, GNN, and our proposed modified algorithm, PROXI-GNN, on the Cora dataset, employing the before mentioned GNN models. The visualization illustrates the performance comparison across different GNN models, highlighting the impact of integrating our edge feature vectors into the models.
Figure 1: Homophilic Setting. Performance comparison of three GNN models (GCN, SAGE, GAT) using the original algorithm (orange) and integrating PROXI vectors via MLP (blue) with the same hyperparameters on CORA dataset.
From the depicted results, it is evident that incorporating our edge feature vectors yields notable improvements in performance across all GNN models examined. Specifically, our modified algorithm PROXI-GNN consistently outperforms the standard algorithm by a margin of at least
Figure 2: Heterophilic Setting. Performance comparison of three GNN models (GCN, SAGE, GAT) using the original algorithm (orange) and integrating PROXI vectors via MLP (blue) the same hyperparameters on TEXAS dataset.
The observed improvements in performance serve as compelling evidence of the efficacy of our proposed approach. By leveraging the additional contextual information embedded in the edge features, our modified algorithm effectively enhances the models' ability to capture intricate relationships and dependencies within the graph structure. This enhancement ultimately leads to more accurate and robust predictions, as demonstrated by the substantial performance gains across the GNN models evaluated.
Overall, the outcomes shown in Figures 1 and 2 highlight the significance and effectiveness of the modifications we suggested. The noteworthy enhancements in performance that may be obtained by using our proposed edge feature vectors underscore the possibility of our methodology to augment the predictive powers of GNN models, especially in situations where edge-level data is pivotal to the underlying graph structure.
In Table 1, we present a performance comparison of GNN models with PROXI-GNN on two distinct datasets, CORA (homophilic) and TEXAS (heterophilic). Notably, the inclusion of our edge features demonstrates significant improvements in classification AUC, i.e., up to 8% in CORA, and 11% in TEXAS. This substantial increase in AUC suggests that the incorporation of our edge features enhances the model's ability to learn from graph structures. Similar results are observed on the heterophilic dataset TEXAS. These results underscore the importance of considering meaningful edge features in graph-based learning tasks, showcasing the efficacy of our PROXI method in improving GNN performance.
Dataset | Backbone | GNN | PROXI-GNN |
---|---|---|---|
CORA | GCN | ||
GAT | |||
SAGE | |||
TEXAS | GCN | ||
GAT | |||
SAGE |