bibtex_url
null | proceedings
stringlengths 42
42
| bibtext
stringlengths 197
848
| abstract
stringlengths 303
3.45k
| title
stringlengths 10
159
| authors
sequencelengths 1
34
⌀ | id
stringclasses 44
values | arxiv_id
stringlengths 0
10
| GitHub
sequencelengths 1
1
| paper_page
stringclasses 899
values | n_linked_authors
int64 -1
13
| upvotes
int64 -1
109
| num_comments
int64 -1
13
| n_authors
int64 -1
92
| Models
sequencelengths 0
100
| Datasets
sequencelengths 0
19
| Spaces
sequencelengths 0
100
| old_Models
sequencelengths 0
100
| old_Datasets
sequencelengths 0
19
| old_Spaces
sequencelengths 0
100
| paper_page_exists_pre_conf
int64 0
1
| type
stringclasses 2
values |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | https://openreview.net/forum?id=shwYa41B3G | @inproceedings{
liao2023gentkg,
title={Gen{TKG}: Generative Forecasting on Temporal Knowledge Graph},
author={Ruotong Liao and Xu Jia and Yunpu Ma and Volker Tresp},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=shwYa41B3G}
} | The rapid advancements in large language models (LLMs) have ignited interest in the realm of the temporal knowledge graph (TKG) domain, where conventional carefully designed embedding-based and rule-based models dominate. The question remains open of whether pre-trained LLMs can understand structured temporal relational data and replace them as the foundation model for temporal relational forecasting. Besides, challenges occur in the huge chasms between complex graph data structure and linear natural expressions LLMs can handle, and between the enormous data volume of TKGs and heavy computation costs of finetuning LLMs. To address these challenges, we bring temporal knowledge forecasting into the generative setting and propose a novel retrieval augmented generation framework named GenTKG combining a temporal logical rule-based retrieval strategy and lightweight few-shot parameter-efficient instruction tuning to solve the above challenges. Extensive experiments have shown that GenTKG is a simple but effective, efficient, and generalizable approach that outperforms conventional methods on temporal relational forecasting with extremely limited computation. Our work opens a new frontier for the temporal knowledge graph domain. | GenTKG: Generative Forecasting on Temporal Knowledge Graph | [
"Ruotong Liao",
"Xu Jia",
"Yunpu Ma",
"Volker Tresp"
] | Workshop/GLFrontiers | [
"https://github.com/mayhugotong/gentkg"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=shKEocsDbB | @inproceedings{
tsitsulin2023the,
title={The Graph Lottery Ticket Hypothesis: Finding Sparse, Informative Graph Structure},
author={Anton Tsitsulin and Bryan Perozzi},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=shKEocsDbB}
} | Graph learning methods help utilize implicit relationships among data items, thereby reducing training label requirements and improving task performance. However, determining the optimal graph structure for a particular learning task remains a challenging research problem.
In this work, we introduce the Graph Lottery Ticket (GLT) Hypothesis – that there is an extremely sparse backbone for every graph, and that graph learning algorithms attain comparable performance when trained on that subgraph as on the full graph.
We identify and systematically study 8 key metrics of interest that directly influence the performance of graph learning algorithms.
Subsequently, we define the notion of a "winning ticket" for graph structure – an extremely sparse subset of edges that can deliver a robust approximation of the entire graph's performance. We propose a straightforward and efficient algorithm for finding these GLTs in arbitrary graphs. Empirically, we observe that performance of different graph learning algorithms can be matched or even exceeded on graphs with the average degree as low as 5. | The Graph Lottery Ticket Hypothesis: Finding Sparse, Informative Graph Structure | [
"Anton Tsitsulin",
"Bryan Perozzi"
] | Workshop/GLFrontiers | 2312.04762 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=sCk5NkZ8Ow | @inproceedings{
choi2023node,
title={Node Mutual Information: Enhancing Graph Neural Networks for Heterophily},
author={Seongjin Choi and Gahee Kim and Se-Young Yun},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=sCk5NkZ8Ow}
} | Graph neural networks (GNNs) have achieved great success in graph analysis by leveraging homophily, where connected nodes share similar properties. However, GNNs struggle on heterophilic graphs where connected nodes tend to differ. Some of the existing methods use neighborhood expansion which is intractable for large graphs.
This paper proposes utilizing node mutual information (MI) to capture dependencies between nodes in heterophilic graphs for use in GNNs. We first define a probability space associated with the graph and introduce $k^{th}$ node random variables to partition the graph based on node distances. The MI between two nodes' random variables then quantifies their dependency regardless of distance by considering both direct and indirect connections. We propose $k^{th}$ MIGNN where the $k^{th}$ MI values are used as weights in the message aggregation function. Experiments on real-world datasets with varying heterophily ratios show the proposed method achieves competitive performance compared to baseline GNNs. The results demonstrate that leveraging node mutual information effectively captures complex node dependencies in heterophilic graphs. | Node Mutual Information: Enhancing Graph Neural Networks for Heterophily | [
"Seongjin Choi",
"Gahee Kim",
"Se-Young Yun"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=sC6NcC4ns9 | @inproceedings{
shirzad2023lowwidth,
title={Low-Width Approximations and Sparsification for Scaling Graph Transformers},
author={Hamed Shirzad and Balaji Venkatachalam and Ameya Velingker and Danica Sutherland and David Woodruff},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=sC6NcC4ns9}
} | Graph Transformers have shown excellent results on a diverse set of datasets. However, memory limitations prohibit these models from scaling to larger graphs. With standard single-GPU setups, even training on medium-sized graphs is impossible for most Graph Transformers. While the $\mathcal{O}(nd^2+n^2d)$ complexity of each layer can be reduced to $\mathcal{O}((n+m)d+nd^2)$ using sparse attention models such as Exphormer for graphs with $n$ nodes and $m$ edges, these models are still infeasible to train on training on small-memory devices even for medium-sized datasets. Here, we propose to sparsify the Exphormer model even further,
by using a small ``pilot'' network to estimate attention scores along the graph edges,
then training a larger model only using $\mathcal O(n)$ edges deemed important by the small network.
We show empirically that attention scores from smaller networks provide a good estimate of the attention scores in larger networks,
and that this process can yield a large-width sparse model nearly as good as the large-width non-sparse model. | Low-Width Approximations and Sparsification for Scaling Graph Transformers | [
"Hamed Shirzad",
"Balaji Venkatachalam",
"Ameya Velingker",
"Danica Sutherland",
"David Woodruff"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=qYpqrt6X3I | @inproceedings{
teneva2023knowledge,
title={Knowledge Graphs are not Created Equal: Exploring the Properties and Structure of Real {KG}s},
author={Nedelina Teneva and Estevam Hruschka},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=qYpqrt6X3I}
} | Despite the recent popularity of knowledge graph (KG) related tasks and benchmarks such as KG embeddings, link prediction, entity alignment and evaluation of the reasoning abilities of pretrained language models as KGs, the structure and properties of real KGs are not well studied. In this paper, we perform a large scale comparative study of 29 real KG datasets from diverse domains such as the natural sciences, medicine, and NLP to analyze their properties and structural patterns. Based on our findings, we make several recommendations regarding KG-based model development and evaluation. We believe that the rich structural information contained in KGs can benefit the development of better KG models across fields and we hope this study will contribute to breaking the existing data silos between different areas of research (e.g., ML, NLP, AI for sciences). | Knowledge Graphs are not Created Equal: Exploring the Properties and Structure of Real KGs | [
"Nedelina Teneva",
"Estevam Hruschka"
] | Workshop/GLFrontiers | 2311.06414 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=qCgptbOsdS | @inproceedings{
pan2023fedgkd,
title={Fed{GKD}: Unleashing the Power of Collaboration in Federated Graph Neural Networks},
author={Qiying Pan and Ruofan Wu and Tengfei LIU and Tianyi Zhang and Yifei Zhu and Weiqiang Wang},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=qCgptbOsdS}
} | Federated training of Graph Neural Networks (GNN) has become popular in recent years due to its ability to perform graph-related tasks under data isolation scenarios while preserving data privacy. However, graph heterogeneity issues in federated GNN systems continue to pose challenges. Existing frameworks address the problem by representing local tasks using different statistics and relating them through a simple aggregation mechanism. However, these approaches suffer from limited efficiency from two aspects: low quality of task-relatedness quantification and inefficacy of exploiting the collaboration structure. To address these issues, we propose FedGKD, a novel federated GNN framework that utilizes a novel client-side graph dataset distillation method to extract task features that better describe task-relatedness, and introduces a novel server-side aggregation mechanism that is aware of the global collaboration structure. We conduct extensive experiments on six real-world datasets of different scales, demonstrating our framework's outperformance. | FedGKD: Unleashing the Power of Collaboration in Federated Graph Neural Networks | [
"Qiying Pan",
"Ruofan Wu",
"Tengfei LIU",
"Tianyi Zhang",
"Yifei Zhu",
"Weiqiang Wang"
] | Workshop/GLFrontiers | 2309.09517 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=q2xXh4M9Dx | @inproceedings{
graziani2023no,
title={No {PAIN} no Gain: More Expressive {GNN}s with Paths},
author={Caterina Graziani and Tamara Drucks and Monica Bianchini and franco scarselli and Thomas G{\"a}rtner},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=q2xXh4M9Dx}
} | Motivated by the lack of theoretical investigation into the discriminative power of _paths_, we characterize classes of graphs where paths are sufficient to identify every instance. Our analysis motivates the integration of paths into the learning procedure of graph neural networks in order to enhance their expressiveness. We formally justify the use of paths based on finite-variable counting logic and prove the effectiveness of paths to recognize graph structural features related to cycles and connectivity. We show that paths are able to identify graphs for which higher-order models fail. Building on this, we propose PAth Isomorphism Network (PAIN), a novel graph neural network that replaces the topological neighborhood with paths in the aggregation step of the message-passing procedure. This modification leads to an algorithm that is strictly more expressive than the Weisfeiler-Leman graph isomorphism test, at the cost of a polynomial-time step for every iteration and fixed path length. We support our theoretical findings by empirically evaluating PAIN on synthetic datasets. | No PAIN no Gain: More Expressive GNNs with Paths | [
"Caterina Graziani",
"Tamara Drucks",
"Monica Bianchini",
"franco scarselli",
"Thomas Gärtner"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=nMdpkeYKzJ | @inproceedings{
erdogan2023poisoning,
title={Poisoning \${\textbackslash}times\$ Evasion: Symbiotic Adversarial Robustness for Graph Neural Networks},
author={Ege Erdogan and Simon Geisler and Stephan G{\"u}nnemann},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=nMdpkeYKzJ}
} | It is well-known that deep learning models are vulnerable w.r.t. small input perturbations. Such perturbed instances are called adversarial examples. Adversarial examples are commonly crafted to fool a model either at training time (poisoning) or test time (evasion). In this work, we study the symbiosis of poisoning and evasion. We show that combining both threat models can substantially improve the devastating efficacy of adversarial attacks. Specifically, we study the robustness of Graph Neural Networks (GNNs) under structure perturbations and devise a memory-efficient adaptive end-to-end attack for the novel threat model using first-order optimization. | Poisoning × Evasion: Symbiotic Adversarial Robustness for Graph Neural Networks | [
"Ege Erdogan",
"Simon Geisler",
"Stephan Günnemann"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=mv5sZrQwLM | @inproceedings{
cohen-karlik2023order,
title={Order Agnostic Autoregressive Graph Generation},
author={Edo Cohen-Karlik and Eyal Rozenberg and Daniel Freedman},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=mv5sZrQwLM}
} | Graph generation is a fundamental problem in various domains, including chemistry and social networks. Recent work has shown that molecular graph generation using recurrent neural networks (RNNs) is advantageous compared to traditional generative approaches which require converting continuous latent representations into graphs. One issue which arises when treating graph generation as sequential generation is the arbitrary order of the sequence which results from a particular choice of graph flattening method. In this work we propose using RNNs, taking into account the non-sequential nature of graphs by adding an Orderless Regularization (OLR) term that encourages the hidden state of the recurrent model to be invariant to different valid orderings present under the training distribution. We demonstrate that sequential graph generation models benefit from our proposed regularization scheme, especially when data is scarce. Our findings contribute to the growing body of research on graph generation and provide a valuable tool for various applications requiring the synthesis of realistic and diverse graph structures. | Order Agnostic Autoregressive Graph Generation | [
"Edo Cohen-Karlik",
"Eyal Rozenberg",
"Daniel Freedman"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=mstzBSOx2e | @inproceedings{
ravichandran2023graphrnn,
title={Graph{RNN} Revisited: An Ablation Study and Extensions for Directed Acyclic Graphs},
author={Maya Ravichandran and Mark Koch and Taniya Das and Nikhil Khatri},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=mstzBSOx2e}
} | GraphRNN is a deep learning-based architecture proposed by You et al. for learning generative models for graphs. We replicate the results of You et al. using a reproduced implementation of the GraphRNN architecture and evaluate this against baseline models using new metrics. Through an ablation study, we find that the BFS traversal suggested by You et al. to collapse representations of isomorphic graphs contributes significantly to model performance. Additionally, we extend GraphRNN to generate directed acyclic graphs by replacing the BFS traversal with a topological sort. We demonstrate that this method improves significantly over a directed-multiclass variant of GraphRNN on a real-world dataset. | GraphRNN Revisited: An Ablation Study and Extensions for Directed Acyclic Graphs | [
"Maya Ravichandran",
"Mark Koch",
"Taniya Das",
"Nikhil Khatri"
] | Workshop/GLFrontiers | 2307.14109 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=m4mF2T0bVm | @inproceedings{
wang2023knowledge,
title={Knowledge Graph Prompting for Multi-Document Question Answering},
author={Yu Wang and Nedim Lipka and Ryan Rossi and Alexa Siu and Ruiyi Zhang and Tyler Derr},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=m4mF2T0bVm}
} | The 'pre-train, prompt, predict' paradigm of large language models (LLMs) has achieved remarkable success in open-domain question answering (OD-QA). However, few works explore this paradigm in the scenario of multi-document question answering (MD-QA), a task demanding a thorough understanding of the logical associations among the contents and structures of different documents. To fill this crucial gap, we propose a Knowledge Graph Prompting (KGP) method to formulate the right context in prompting LLMs for MD-QA, which consists of a graph construction module and a graph traversal module. For graph construction, we create a knowledge graph (KG) over multiple documents with nodes symbolizing passages or document structures (e.g., pages/tables), and edges denoting the semantic/lexical similarity between passages or intra-document structural relations. For graph traversal, we design an LM-guided graph traverser that navigates across nodes and gathers supporting passages assisting LLMs in MD-QA. The constructed graph serves as the global ruler that regulates the transitional space among passages and reduces retrieval latency. Concurrently, the LM-guided traverser acts as a local navigator that gathers pertinent context to progressively approach the question and guarantee retrieval quality. Extensive experiments underscore the efficacy of KGP for MD-QA, signifying the potential of leveraging graphs in enhancing the prompt design for LLMs. Our code will be released upon publication. | Knowledge Graph Prompting for Multi-Document Question Answering | [
"Yu Wang",
"Nedim Lipka",
"Ryan Rossi",
"Alexa Siu",
"Ruiyi Zhang",
"Tyler Derr"
] | Workshop/GLFrontiers | 2308.11730 | [
"https://github.com/yuwvandy/kg-llm-mdqa"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=m1WEVCY1O2 | @inproceedings{
zuo2023dipgnn,
title={DiP-{GNN}: Discriminative Pre-Training of Graph Neural Networks},
author={Simiao Zuo and Haoming Jiang and Qingyu Yin and Xianfeng Tang and Bing Yin and Tuo Zhao},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=m1WEVCY1O2}
} | Graph neural network (GNN) pre-training methods have been proposed to enhance the power of GNNs. Specifically, a GNN is first pre-trained on a large-scale unlabeled graph and then fine-tuned on a separate small labeled graph for downstream applications, such as node classification. One popular pre-training method is to mask out a proportion of the edges, and a GNN is trained to recover them. However, such a generative method suffers from graph mismatch. That is, the masked graph input to the GNN deviates from the original graph. To alleviate this issue, we propose DiP-GNN (Discriminative Pre-training of Graph Neural Networks). Specifically, we train a generator to recover identities of the masked edges, and simultaneously, we train a discriminator to distinguish the generated edges from the original graph's edges. The discriminator is subsequently used for downstream fine-tuning. In our pre-training framework, the graph seen by the discriminator better matches the original graph because the generator can recover a proportion of the masked edges. Extensive experiments on large-scale homogeneous and heterogeneous graphs demonstrate the effectiveness of DiP-GNN. Our code will be publicly available. | DiP-GNN: Discriminative Pre-Training of Graph Neural Networks | [
"Simiao Zuo",
"Haoming Jiang",
"Qingyu Yin",
"Xianfeng Tang",
"Bing Yin",
"Tuo Zhao"
] | Workshop/GLFrontiers | 2209.07499 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=lSa6SEEqTL | @inproceedings{
kang2023coupling,
title={Coupling Graph Neural Networks with Non-Integer Order Dynamics: A Robustness Study},
author={Qiyu Kang and Kai Zhao and Yang Song and Yihang Xie and Yanan Zhao and Sijie Wang and Rui She and Wee Peng Tay},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=lSa6SEEqTL}
} | In this work, we rigorously investigate the robustness of graph neural fractional-order differential equation (FDE) models. This framework extends beyond traditional graph neural ordinary differential equation (ODE) models by implementing the time-fractional Caputo derivative. Utilizing fractional calculus allows our model to consider long-term dependencies during the feature updating process, diverging from the Markovian updates seen in traditional graph neural ODE models.
The efficacy of FDE models in surpassing ODE models has been confirmed in a different submitted work, particularly in environments free from attacks or perturbations.
While traditional graph neural ODE models have been verified to possess a degree of stability and resilience in the presence of adversarial attacks in existing literature, the robustness of graph neural FDE models, especially under adversarial conditions, remains largely unexplored. This paper undertakes a detailed assessment of the robustness of graph neural FDE models. We establish a theoretical foundation outlining the robustness features of graph neural FDE models, highlighting that they maintain more stringent output perturbation bounds in the face of input and functional disturbances, relative to their integer-order counterparts. Through rigorous experimental assessments, which include graph alteration scenarios and adversarial attack contexts, we empirically validate the improved robustness of graph neural FDE models against their conventional graph neural ODE counterparts. | Coupling Graph Neural Networks with Non-Integer Order Dynamics: A Robustness Study | [
"Qiyu Kang",
"Kai Zhao",
"Yang Song",
"Yihang Xie",
"Yanan Zhao",
"Sijie Wang",
"Rui She",
"Wee Peng Tay"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=lR5NYB9zrv | @inproceedings{
lachi2023graph,
title={Graph Pooling Provably Improves Expressivity},
author={Veronica Lachi and Alice Moallemy-Oureh and Andreas Roth and Pascal Welke},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=lR5NYB9zrv}
} | In the domain of graph neural networks (GNNs), pooling operators are fundamental to reduce the size of the graph by simplifying graph structures and vertex features.
Recent advances have shown that well-designed pooling operators, coupled with message-passing layers, can endow hierarchical GNNs with an expressive power regarding the graph isomorphism test that is equal to the Weisfeiler-Leman test.
However, the ability of hierarchical GNNs to increase expressive power by utilizing graph coarsening was not yet explored.
This results in uncertainties about the benefits of pooling operators and a lack of sufficient properties to guide their design.
In this work, we identify conditions for pooling operators to generate WL-*distinguishable* coarsened graphs from originally WL-*indistinguishable* but non-isomorphic graphs.
Our conditions are versatile and can be tailored to specific tasks and data characteristics, offering a promising avenue for further research. | Graph Pooling Provably Improves Expressivity | [
"Veronica Lachi",
"Alice Moallemy-Oureh",
"Andreas Roth",
"Pascal Welke"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=jyfiPivRBH | @inproceedings{
huang2023can,
title={Can {LLM}s Effectively Leverage Graph Structural Information: When and Why},
author={Jin Huang and Xingjian Zhang and Qiaozhu Mei and Jiaqi Ma},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=jyfiPivRBH}
} | This paper studies Large Language Models (LLMs) augmented with structured data--particularly graphs--a crucial data modality that remains underexplored in the LLM literature. We aim to understand when and why the incorporation of structural information inherent in graph data can improve the prediction performance of LLMs on node classification tasks with textual features. To address the "when" question, we examine a variety of prompting methods for encoding structural information, in settings where textual node features are either rich or scarce. For the "why" questions, we probe into two potential contributing factors to the LLM performance: data leakage and homophily. Our exploration of these questions reveals that (i) LLMs can benefit from structural information, especially when textual node features are scarce; (ii) there is no substantial evidence indicating that the performance of LLMs is significantly attributed to data leakage; and (iii) the performance of LLMs on a target node is strongly positively related to the local homophily ratio of the node. | Can LLMs Effectively Leverage Graph Structural Information: When and Why | [
"Jin Huang",
"Xingjian Zhang",
"Qiaozhu Mei",
"Jiaqi Ma"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=jbx2WMP9VX | @inproceedings{
liu2023semisupervised,
title={Semi-Supervised Graph Imbalanced Regression},
author={Gang Liu and Tong Zhao and Eric Inae and Tengfei Luo and Meng Jiang},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=jbx2WMP9VX}
} | Data imbalance is easily found in annotated data when the observations of certain continuous label values are difficult to collect for regression tasks. When they come to molecule and polymer property predictions, the annotated graph datasets are often small because labeling them requires expensive equipment and effort. To address the lack of examples of rare label values in graph regression tasks, we propose a semi-supervised framework to progressively balance training data and reduce model bias via self-training. The training data balance is achieved by (1) pseudo-labeling more graphs for under-represented labels with a novel regression confidence measurement and (2) augmenting graph examples in latent space for remaining rare labels after data balancing with pseudo-labels. The former is to identify quality examples from unlabeled data whose labels are confidently predicted and sample a subset of them with a reverse distribution from the imbalanced annotated data. The latter collaborates with the former to target a perfect balance using a novel label-anchored mixup algorithm. We perform experiments in seven regression tasks on graph datasets. Results demonstrate that the proposed framework significantly reduces the error of predicted graph properties, especially in under-represented label areas. | Semi-Supervised Graph Imbalanced Regression | [
"Gang Liu",
"Tong Zhao",
"Eric Inae",
"Tengfei Luo",
"Meng Jiang"
] | Workshop/GLFrontiers | 2305.12087 | [
"https://github.com/liugangcode/SGIR"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=jNL90brxG0 | @inproceedings{
hu2023beyond,
title={Beyond Text: A Deep Dive into Large Language Models' Ability on Understanding Graph Data},
author={Yuntong Hu and Zheng Zhang and Liang Zhao},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=jNL90brxG0}
} | Large language models (LLMs) have achieved impressive performance on many natural language processing tasks. However, their capabilities on graph-structured data remain relatively unexplored. In this paper, we conduct a series of experiments benchmarking leading LLMs on diverse graph prediction tasks spanning node, edge, and graph levels. We aim to assess whether LLMs can effectively process graph data and leverage topological structures to enhance performance, compared to specialized graph neural networks. Through varied prompt formatting and task/dataset selection, we analyze how well LLMs can interpret and utilize graph structures. By comparing LLMs' performance with specialized graph models, we offer insights into the strengths and limitations of employing LLMs for graph analytics. Our findings provide insights into LLMs' capabilities and suggest avenues for further exploration in applying them to graph analytics. | Beyond Text: A Deep Dive into Large Language Models' Ability on Understanding Graph Data | [
"Yuntong Hu",
"Zheng Zhang",
"Liang Zhao"
] | Workshop/GLFrontiers | 2310.04944 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=i8JgQVS30K | @inproceedings{
davies2023its,
title={Its All Graph To Me: Single-Model Graph Representation Learning on Multiple Domains},
author={Alex Davies and Riku Green and Nirav Ajmeri and Telmo Silva Filho},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=i8JgQVS30K}
} | Graph neural networks (GNNs) have revolutionised the field of graph representation learning and plays a critical role in graph-based research. Recent work explores applying GNNs to pre-training and fine-tuning, where a model is trained on a large dataset and its learnt representations are then transferred to a smaller dataset. However, current work only explore pre-training on a single domain; for example, a model pre-trained on molecular graphs is fine-tuned on other molecular graphs. This leads to poor generalisability of pre-trained models to novel domains and tasks.
In this work, we curate a multi-graph-domain dataset and apply state-of-the-art Graph Adversarial Contrastive Learning (GACL) methods. We present a pre-trained graph model that may have the capability of acting as a foundational graph model. We will evaluate the efficacy of its learnt representations on various downstream tasks against baseline models pre-trained on single domains. In addition, we aim to compare our model to un-trained and non-transferred models, and show that performance using our foundational model is capable of achieving equal or better than task-specific methodology. | Its All Graph To Me: Single-Model Graph Representation Learning on Multiple Domains | [
"Alex Davies",
"Riku Green",
"Nirav Ajmeri",
"Telmo Silva Filho"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=hySoEBuTuM | @inproceedings{
li2023longrange,
title={Long-Range Neural Atom Learning for Molecular Graphs},
author={Xuan Li and Zhanke Zhou and Jiangchao Yao and Yu Rong and Lu Zhang and Bo Han},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=hySoEBuTuM}
} | Graph Neural Networks (GNNs) have been widely adopted for drug discovery with molecular graphs. Nevertheless, current GNNs are mainly good at leveraging short-range interactions (SRI) but struggle to capture long-range interactions (LRI), both of which are crucial for determining molecular properties. To tackle this issue, we propose a method that implicitly projects all original atoms into a few \textit{Neural Atoms}, which abstracts the collective information of atomic groups within a molecule. Specifically, we explicitly exchange the information among neural atoms and project them back to the atoms’ representations as an enhancement. With this mechanism, neural atoms establish the communication channels among distant nodes, effectively reducing the interaction scope of arbitrary node pairs into a single hop. To provide an inspection of our method from a physical perspective, we reveal its connection with the traditional LRI calculation method, Ewald Summation. We conduct extensive experiments on three long-range graph benchmarks, covering both graph-level and link-level tasks on molecular graphs. We empirically justify that our method can be equipped with an arbitrary GNN and help to capture LRI. | Long-Range Neural Atom Learning for Molecular Graphs | [
"Xuan Li",
"Zhanke Zhou",
"Jiangchao Yao",
"Yu Rong",
"Lu Zhang",
"Bo Han"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=gXCqzpvhuD | @inproceedings{
bause2023maximally,
title={Maximally Expressive {GNN}s for Outerplanar Graphs},
author={Franka Bause and Fabian Jogl and Patrick Indri and Tamara Drucks and David Penz and Nils Kriege and Thomas G{\"a}rtner and Pascal Welke and Maximilian Thiessen},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=gXCqzpvhuD}
} | We propose a linear time graph transformation that enables the Weisfeiler-Leman (WL) test and message passing graph neural networks (MPNNs) to be maximally expressive on outerplanar graphs. Our approach is motivated by the fact that most pharmaceutical molecules correspond to outerplanar graphs. Existing research predominantly enhances the expressivity of graph neural networks without specific graph families in mind. This often leads to methods that are impractical due to their computational complexity. In contrast, the restriction to outerplanar graphs enables us to encode the Hamiltonian cycle of each biconnected component in linear time. As the main contribution of the paper we prove that our method achieves maximum expressivity on outerplanar graphs. Experiments confirm that our graph transformation improves the predictive performance of MPNNs on molecular benchmark datasets at negligible computational overhead. | Maximally Expressive GNNs for Outerplanar Graphs | [
"Franka Bause",
"Fabian Jogl",
"Patrick Indri",
"Tamara Drucks",
"David Penz",
"Nils Kriege",
"Thomas Gärtner",
"Pascal Welke",
"Maximilian Thiessen"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
||
null | https://openreview.net/forum?id=gVn2F8MBnP | @inproceedings{
m{\"u}ller2023privacyutility,
title={Privacy-Utility Trade-offs in Neural Networks for Medical Population Graphs: Insights from Differential Privacy and Graph Structure},
author={Tamara M{\"u}ller and Maulik Chevli and Ameya Daigavane and Daniel Rueckert and Georgios Kaissis},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=gVn2F8MBnP}
} | Differential privacy (DP) is the gold standard for protecting individuals' data while enabling deep learning. It is well-established and frequently used for applications in medicine and healthcare to protect sensitive patient data. When using graph deep learning on so-called population graphs, however, the application of DP becomes more challenging compared to grid-like data structures like images or tables. In this work, we initiate an empirical investigation of differentially private graph neural networks on population graphs in the medical domain by examining privacy-utility trade-offs under different graph learning methods on both real-world and synthetic datasets. We compare two state-of-the-art methods for differentially private graph deep learning and empirically audit privacy guarantees through node membership inference and link stealing attacks. We focus on the impact of the graph structure, one of the most important inherent challenges of medical population graphs. Our findings highlight the potential and challenges of this specific DP application area. Moreover, we find that the underlying graph structure constitutes a potential factor for larger performance gaps on one of the explored methods by showing a correlation between the graphs' homophily and the accuracy of the trained model. | Privacy-Utility Trade-offs in Neural Networks for Medical Population Graphs: Insights from Differential Privacy and Graph Structure | [
"Tamara Müller",
"Maulik Chevli",
"Ameya Daigavane",
"Daniel Rueckert",
"Georgios Kaissis"
] | Workshop/GLFrontiers | 2307.06760 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=fLEPS2BwRL | @inproceedings{
behrouz2023unsupervised,
title={Unsupervised Representation Learning of Brain Activity via Bridging Voxel Activity and Functional Connectivity},
author={Ali Behrouz and Parsa Delavari and Farnoosh Hashemi},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=fLEPS2BwRL}
} | Effective brain representation learning is a key step toward revealing the understanding of cognitive processes and unlocking detecting and potential therapeutic interventions for neurological diseases/disorders. Existing studies have focused on either (1) voxel-level activity, where only a single beta weight for each voxel (i.e., aggregation of voxel activity over a time window) is considered, missing their temporal dynamics, or (2) functional connectivity of the brain in the level of region of interests, missing voxel-level activities. In this paper, we bridge this gap and design BrainMixer, an unsupervised learning framework that effectively utilizes both functional connectivity and associated time series of voxels to learn voxel-level representation in an unsupervised manner. BrainMixer employs two simple yet effective MLP-based encoders to simultaneously learn the dynamics of voxel-level signals and their functional correlations. To encode voxel activity, BrainMixer fuses information across both time and voxel dimensions via a dynamic self-attention mechanism. To learn the structure of the functional connectivity graph, BrainMixer presents a temporal graph patching and encodes each patch by combining its nodes' features via a new adaptive temporal pooling. Our experiments show that BrainMixer attains outstanding performance and outperforms 13 baselines in different downstream tasks and experimental setups. | Unsupervised Representation Learning of Brain Activity via Bridging Voxel Activity and Functional Connectivity | [
"Ali Behrouz",
"Parsa Delavari",
"Farnoosh Hashemi"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=esQEXmBqxR | @inproceedings{
koke2023holonets,
title={HoloNets: Spectral Convolutions do extend to Directed Graphs},
author={Christian Koke and Daniel Cremers},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=esQEXmBqxR}
} | Within the graph learning community, conventional wisdom dictates that spectral convolutional networks may only be deployed on undirected graphs: Only there could the existence of a well-defined graph Fourier transform be guaranteed, so that information may be translated between spatial- and spectral domains. Here we show this traditional reliance on the graph Fourier transform to be superfluous: Making use of certain advanced tools from complex analysis and spectral theory, we extend spectral convolutions to directed graphs. We provide a frequency- response interpretation of newly developed filters, investigate the influence of the basis’ used to express filters and discuss the interplay with characteristic operators on which networks are based. In order to thoroughly test the developed general theory, we conduct experiments in real world settings, showcasing that directed spectral convolutional networks provide new state of the art results for heterophilic node classification and – as opposed to baselines – may be rendered stable to resolution-scale varying topological perturbations. | HoloNets: Spectral Convolutions do extend to Directed Graphs | [
"Christian Koke",
"Daniel Cremers"
] | Workshop/GLFrontiers | 2310.02232 | [
"https://github.com/ChristianKoke/HoloNets"
] | https://huggingface.co/papers/2310.02232 | 0 | 0 | 0 | 2 | [] | [] | [] | [] | [] | [] | 1 | oral |
null | https://openreview.net/forum?id=eYyZisiBG3 | @inproceedings{
cavazos2023explaining,
title={Explaining Drug Repositioning: A Case-Based Reasoning Graph Neural Network Approach},
author={Adriana Gonzalez Cavazos},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=eYyZisiBG3}
} | Drug repositioning, the identification of novel uses of existing therapies, has become an attractive strategy to accelerate drug development. Knowledge graphs (KGs) have emerged as a powerful representation of interconnected data within the biomedical domain. While link prediction on biomedical can ascertain new connections between drugs and diseases, most approaches only state whether two
nodes are related. Yet, they fail to explain why two nodes are related. In this project, we introduce an implementation of the semi-parametric Case-Based Reasoning over subgraphs (CBR-SUBG), designed to derive a drug query’s underlying mechanisms by gathering graph patterns of similar nodes. We show that our adaptation outperforms existing KG link prediction models on a drug repositioning task.
Furthermore, our findings demonstrate that CBR-SUBG strategy can provide interpretable biological paths as evidence supporting putative repositioning candidates, leading to more informed decisions. | Explaining Drug Repositioning: A Case-Based Reasoning Graph Neural Network Approach | [
"Adriana Gonzalez Cavazos"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=e8ba9Hu1mM | @inproceedings{
bar-shalom2023subgraphormer,
title={Subgraphormer: Subgraph {GNN}s meet Graph Transformers},
author={Guy Bar-Shalom and Beatrice Bevilacqua and Haggai Maron},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=e8ba9Hu1mM}
} | In the realm of Graph Neural Network (GNNs), two intriguing research directions have recently emerged: Subgraph GNNs and Graph Transformers. These approaches have distinct origins -- Subgraph GNNs aim to address the limitations of message passing, while Graph Transformers seek to build on the success of sequential transformers in language and vision tasks. In this paper, we propose a model that integrates both approaches, dubbed _Subgraphormer_, which combines the message passing and global aggregation schemes from Subgraph GNNs with attention mechanisms and positional and structural encodings, which are arguably the most important components in Graph Transformers. Our preliminary experimental results demonstrate significant performance improvements over both Subgraph GNNs and Graph Transformers. | Subgraphormer: Subgraph GNNs meet Graph Transformers | [
"Guy Bar-Shalom",
"Beatrice Bevilacqua",
"Haggai Maron"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=dC3CaRZ8tf | @inproceedings{
j{\"u}r{\ss}2023everybody,
title={Everybody Needs a Little {HELP}: Explaining Graphs via Hierarchical Concepts},
author={Jonas J{\"u}r{\ss} and Lucie Charlotte Magister and Pietro Barbiero and Pietro Lio and Nikola Simidjievski},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=dC3CaRZ8tf}
} | Graph neural networks (GNNs) have led to major breakthroughs in a variety of domains such as drug discovery, social network analysis, and travel time estimation. However, they lack interpretability which hinders human trust and thereby deployment to settings with high-stakes decisions. A line of interpretable methods approach this by discovering a small set of relevant concepts as subgraphs in the last GNN layer that together explain the prediction. This can yield oversimplified explanations, failing to explain the interaction between GNN layers. To address this oversight, we provide HELP (Hierarchical Explainable Latent Pooling), a novel, inherently interpretable graph pooling approach that reveals how concepts from different GNN layers compose to new ones in later steps. HELP is more than 1-WL expressive and is the first non-spectral, end-to-end-learnable, hierarchical graph pooling method that can learn to pool a variable number of arbitrary connected components. We empirically demonstrate that it performs on-par with standard GCNs and popular pooling methods in terms of accuracy while yielding explanations that are aligned with expert knowledge in the domains of chemistry and social networks. In addition to a qualitative analysis, we employ concept completeness scores as well as concept conformity, a novel metric to measure the noise in discovered concepts, quantitatively verifying that the discovered concepts are significantly easier to fully understand than those from previous work. Our work represents a first step towards an understanding of graph neural networks that goes beyond a set of concepts from the final layer and instead explains the complex interplay of concepts on different levels. | Everybody Needs a Little HELP: Explaining Graphs via Hierarchical Concepts | [
"Jonas Jürß",
"Lucie Charlotte Magister",
"Pietro Barbiero",
"Pietro Lio",
"Nikola Simidjievski"
] | Workshop/GLFrontiers | 2311.15112 | [
"https://github.com/jonasjuerss/help"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=d1oVSv97ZM | @inproceedings{
wang2023protohg,
title={Proto{HG}: Prototype-Enhanced Hypergraph Learning for Heterogeneous Information Networks},
author={Shuai Wang and Jiayi Shen and Athanasios Efthymiou and Stevan Rudinac and Monika Kackovic and Nachoem Wijnberg and Marcel Worring},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=d1oVSv97ZM}
} | The variety and complexity of relations in real-world data lead to Heterogeneous Information Networks (HINs). Capturing the semantics from such networks requires approaches capable of utilizing the full richness of the HINs. Existing methods for modeling HINs employ techniques originally designed for graph neural networks in combination with HIN decomposition analysis, especially using manually predefined metapaths. In this paper, we introduce a novel hypergraph learning approach for node classification in HINs. Using hypergraphs instead of graphs, our method captures higher-order relationships among heterogeneous nodes and extracts semantic information without relying on metapaths. Our method leverages the power of prototypes to improve the robustness of the hypergraph learning process, and we further discuss the advantages that our method can bring to scalability, which due to their expressiveness is an important issue for hypergraphs. Extensive experiments on three real-world HINs demonstrate the effectiveness of our method. | ProtoHG: Prototype-Enhanced Hypergraph Learning for Heterogeneous Information Networks | [
"Shuai Wang",
"Jiayi Shen",
"Athanasios Efthymiou",
"Stevan Rudinac",
"Monika Kackovic",
"Nachoem Wijnberg",
"Marcel Worring"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=ck099Dyu4j | @inproceedings{
jin2023learning,
title={Learning Multiplex Embeddings on Text-rich Networks with One Text Encoder},
author={Bowen Jin and Wentao Zhang and Yu Zhang and Yu Meng and Han Zhao and Jiawei Han},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=ck099Dyu4j}
} | In real-world scenarios, texts in a network are often linked by multiple semantic relations (e.g., papers in an academic network are referenced by other publications, written by the same author, or published in the same venue). Text documents and their relations form a multiplex text-rich network. Mainstream text representation learning methods use pretrained language models (PLMs) to generate one embedding for each text unit, expecting that all types of relations between texts can be captured by these single-view embeddings. However, this presumption does not hold particularly in multiplex text-rich networks. Along another line of work, multiplex graph neural networks (GNNs) directly initialize node attributes as a feature vector for node representation learning, but they cannot fully capture the semantics of the nodes' associated texts. In this paper, we propose METERN, a new framework for learning Multiplex Embeddings on TExt-Rich Networks. In contrast to existing methods, METERN uses one text encoder to model the shared knowledge across relations and leverages a small number of parameters per relation to derive relation-specific representations. This allows the encoder to effectively capture the multiplex structures in the network while also preserving parameter efficiency. We conduct experiments on nine downstream tasks in five networks from both academic and e-commerce domains, where METERN outperforms baselines significantly and consistently. The code is available at https://anonymous.4open.science/r/METER-submit-2C7B. | Learning Multiplex Embeddings on Text-rich Networks with One Text Encoder | [
"Bowen Jin",
"Wentao Zhang",
"Yu Zhang",
"Yu Meng",
"Han Zhao",
"Jiawei Han"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=cZMdiwFjCh | @inproceedings{
park2023nonbacktracking,
title={Non-backtracking Graph Neural Networks},
author={Seonghyun Park and Narae Ryu and Gahee Kim and Dongyeop Woo and Se-Young Yun and Sungsoo Ahn},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=cZMdiwFjCh}
} | The celebrated message-passing updates for graph neural networks allow the representation of large-scale graphs with local and computationally tractable updates. However, the local updates suffer from backtracking, i.e., a message flows through the same edge twice and revisits the previously visited node. Since the number of message flows increases exponentially with the number of updates, the redundancy in local updates prevents the graph neural network from accurately recognizing a particular message flow for downstream tasks. In this work, we propose to resolve such a redundancy via the non-backtracking graph neural network (NBA-GNN) that updates a message without incorporating the message from the previously visited node. We further investigate how NBA-GNN alleviates the over-squashing of GNNs, and establish a connection between NBA-GNN and the impressive performance of non-backtracking updates for stochastic block model recovery. We empirically verify the effectiveness of our NBA-GNN on long-range graph benchmark and transductive node classification problems. | Non-backtracking Graph Neural Networks | [
"Seonghyun Park",
"Narae Ryu",
"Gahee Kim",
"Dongyeop Woo",
"Se-Young Yun",
"Sungsoo Ahn"
] | Workshop/GLFrontiers | 2310.07430 | [
"https://github.com/seonghyun26/nba-gnn"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=cVIRwcJ3Cb | @inproceedings{
zhang2023graph,
title={Graph Meets {LLM}s: Towards Large Graph Models},
author={Ziwei Zhang and Haoyang Li and Zeyang Zhang and Yijian Qin and Xin Wang and Wenwu Zhu},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=cVIRwcJ3Cb}
} | Large models have emerged as the most recent groundbreaking achievements in artificial intelligence, and particularly machine learning. However, when it comes to graphs, large models have not achieved the same level of success as in other fields, such as natural language processing and computer vision. In order to promote applying large models for graphs forward, we present a perspective paper to discuss the challenges and opportunities associated with developing large graph models. First, we discuss the desired characteristics of large graph models. Then, we present detailed discussions from three key perspectives: representation basis, graph data, and graph models. In each category, we provide a brief overview of recent advances and highlight the remaining challenges together with our visions. Finally, we discuss valuable applications of large graph models. We believe this perspective can encourage further investigations into large graph models, ultimately pushing us one step closer towards artificial general intelligence (AGI). We are the first to comprehensively study large graph models, to the best of our knowledge. | Graph Meets LLMs: Towards Large Graph Models | [
"Ziwei Zhang",
"Haoyang Li",
"Zeyang Zhang",
"Yijian Qin",
"Xin Wang",
"Wenwu Zhu"
] | Workshop/GLFrontiers | 2308.14522 | [
"https://github.com/thumnlab/awesome-large-graph-model"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=cMz2pdzu8W | @inproceedings{
wu2023privacypreserving,
title={Privacy-preserving design of graph neural networks with applications to vertical federated learning},
author={Ruofan Wu and Mingyang Zhang and Lingjuan Lyu and Xiaolong Xu and Xiuquan Hao and xinyi fu and Tengfei LIU and Tianyi Zhang and Weiqiang Wang},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=cMz2pdzu8W}
} | The paradigm of vertical federated learning (VFL), where institutions collaboratively train machine learning models via combining each other's local feature or label information, has achieved great success in applications to financial risk management (FRM). The surging developments of graph representation learning (GRL) have opened up new opportunities for FRM applications under FL via efficiently utilizing the graph-structured data generated from underlying transaction networks. Meanwhile, transaction information is often considered highly sensitive. To prevent data leakage during training, it is critical to develop FL protocols with **formal privacy guarantees**. In this paper, we present an end-to-end GRL framework in the VFL setting called VESPER, which is built upon a general privatization scheme termed perturbed message passing (PMP) that allows the privatization of many popular graph neural architectures. Based on PMP, we discuss the strengths and weaknesses of specific design choices of concrete graph neural architectures and provide solutions and improvements for both dense and sparse graphs. Extensive empirical evaluations over both public datasets and an industry dataset demonstrate that VESPER is capable of training high-performance GNN models over both sparse and dense graphs under reasonable privacy budgets. | Privacy-preserving design of graph neural networks with applications to vertical federated learning | [
"Ruofan Wu",
"Mingyang Zhang",
"Lingjuan Lyu",
"Xiaolong Xu",
"Xiuquan Hao",
"xinyi fu",
"Tengfei LIU",
"Tianyi Zhang",
"Weiqiang Wang"
] | Workshop/GLFrontiers | 2310.20552 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=bxOD6eMAfa | @inproceedings{
varbella2023powergraph,
title={PowerGraph: A power grid benchmark dataset for graph neural networks},
author={Anna Varbella and Kenza Amara and Blazhe Gjorgiev and Giovanni Sansavini},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=bxOD6eMAfa}
} | Public Graph Neural Networks (GNN) benchmark datasets facilitate the use of GNN and enhance GNN applicability to diverse disciplines. The community currently lacks public datasets of electrical power grids for GNN applications. Indeed, GNNs have the potential for capturing complex power grid phenomena over alternative machine learning techniques. Power grids are complex engineered networks that are naturally amenable to graph representations. Therefore, GNN have the potential for capturing the behaviour of power grids over alternative machine learning techniques. To this aim, we develop a graph dataset for cascading failure events, which are the major cause of blackouts in electric power grids. Historical blackout datasets are scarce and incomplete. The assessment of vulnerability and the identification of critical components are usually conducted via computationally expensive offline simulations of cascading failures. Instead, we propose the use of machine learning models for the online detection of cascading failures leveraging the knowledge of the system state at the onset of the cascade. We develop PowerGraph, a graph dataset modelling cascading failures in power grids, designed for two purposes, namely, i) training GNN models for different graph-level tasks including multi-class classification, binary classification, and regression, and ii) explaining GNN models. The dataset generated via physics-based cascading failure model ensures generality of the operating and environmental conditions by spanning diverse failure scenarios. In addition, we foster the use of the dataset to benchmark GNN explainability methods by assigning ground-truth edge-level explanations. PowerGraph helps the development of better GNN models for graph-level tasks and explainability, critical in many domains ranging from chemistry to biology, where the systems and processes can be described as graphs. The dataset is available at https://figshare.com/articles/dataset/PowerGraph/22820534 and the code at https://anonymous.4open.science/r/PowerGraph/. | PowerGraph: A power grid benchmark dataset for graph neural networks | [
"Anna Varbella",
"Kenza Amara",
"Blazhe Gjorgiev",
"Giovanni Sansavini"
] | Workshop/GLFrontiers | 2402.02827 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=a9loLCbWs1 | @inproceedings{
fu2023implicit,
title={Implicit Graph Neural Diffusion Based on Constrained Dirichlet Energy Minimization},
author={Guoji Fu and Mohammed Haroon Dupty and Yanfei Dong and Wee Sun Lee},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=a9loLCbWs1}
} | Implicit graph neural networks (GNNs) have emerged as a potential approach to enable GNNs to capture long-range dependencies effectively. However, poorly designed implicit GNN layers can experience over-smoothing or may have limited adaptability to learn the graph geometry, potentially hindering their performance in graph learning problems. To address these issues, we introduce a geometric framework to design implicit graph diffusion layers based on a parameterized graph Laplacian operator. Our framework allows learning the metrics of vertex and edge spaces, as well as the graph gradient operator from data. We further show how implicit GNN layers can be viewed as the fixed-point solution of a Dirichlet energy minimization problem and give conditions under which it may suffer from over-smoothing. To overcome the over-smoothing problem, we design our implicit graph diffusion layer as the solution of a Dirichlet energy minimization problem with constraints on vertex features, enabling it to trade off smoothing with the preservation of node feature information. With an appropriate hyperparameter set to be larger than the largest eigenvalue of the parameterized graph Laplacian, our framework guarantees a unique equilibrium and quick convergence. Our models demonstrate better performance than leading implicit and explicit GNNs on benchmark datasets for node and graph classification tasks, with substantial accuracy improvements observed for some datasets. | Implicit Graph Neural Diffusion Based on Constrained Dirichlet Energy Minimization | [
"Guoji Fu",
"Mohammed Haroon Dupty",
"Yanfei Dong",
"Wee Sun Lee"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=YXSJxi8dOV | @inproceedings{
berto2023rlco,
title={{RL}4{CO}: a Unified Reinforcement Learning for Combinatorial Optimization Library},
author={Federico Berto and Chuanbo Hua and Junyoung Park and Minsu Kim and Hyeonah Kim and Jiwoo Son and Haeyeon Kim and Joungho Kim and Jinkyoo Park},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=YXSJxi8dOV}
} | Deep reinforcement learning offers notable benefits in addressing combinatorial problems over traditional solvers, reducing the reliance on domain-specific knowl- edge and expert solutions, and improving computational efficiency. Despite the recent surge in interest in neural combinatorial optimization, practitioners often do not have access to a standardized code base. Moreover, different algorithms are frequently based on fragmentized implementations that hinder reproducibility and fair comparison. To address these challenges, we introduce RL4CO, a uni- fied Reinforcement Learning (RL) for Combinatorial Optimization (CO) library. We employ state-of-the-art software and best practices in implementation, such as modularity and configuration management, to be flexible, easily modifiable, and extensible by researchers. Thanks to our unified codebase, we benchmark baseline RL solvers with different evaluation schemes on zero-shot performance, general- ization, and adaptability on diverse tasks. Notably, we find that some recent meth- ods may fall behind their predecessors depending on the evaluation settings. We hope RL4CO will encourage the exploration of novel solutions to complex real- world tasks, allowing the community to compare with existing methods through a unified framework that decouples the science from software engineering. We open-source our library at https://github.com/ai4co/rl4co. | RL4CO: a Unified Reinforcement Learning for Combinatorial Optimization Library | [
"Federico Berto",
"Chuanbo Hua",
"Junyoung Park",
"Minsu Kim",
"Hyeonah Kim",
"Jiwoo Son",
"Haeyeon Kim",
"Joungho Kim",
"Jinkyoo Park"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
||
null | https://openreview.net/forum?id=YILik4gFBk | @inproceedings{
yoon2023multimodal,
title={Multimodal Graph Learning for Generative Tasks},
author={Minji Yoon and Jing Yu Koh and Bryan Hooi and Russ Salakhutdinov},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=YILik4gFBk}
} | Multimodal learning combines multiple data modalities, broadening the types and complexity of data our models can utilize; for example, from plain text to image-caption pairs. Most multimodal learning algorithms focus on modeling simple one-to-one pairs of data from two modalities, such as image-caption pairs, or audio-text pairs. However, in most real-world settings, entities of different modalities interact with each other in more complex and multifaceted ways, going beyond one-to-one mappings. We propose to represent these complex relationships as graphs, allowing us to capture data with any number of modalities, and with complex relationships between modalities that can flexibly vary from one sample to another. Toward this goal, we propose Multimodal Graph Learning (MMGL), a general and systematic framework for capturing information from multiple multimodal neighbors with relational structures among them. In particular, we focus on MMGL for \emph{generative} tasks, building upon pretrained Language Models (LMs), aiming to augment their text generation with multimodal neighbor contexts. We study three research questions raised by MMGL: (1) how can we infuse multiple neighbor information into the pretrained LMs, while avoiding scalability issues? (2) how can we infuse the graph structure information among multimodal neighbors into the LMs? and (3) how can we finetune the pretrained LMs to learn from the neighbor context parameter-efficiently? We conduct extensive experiments to answer these three questions on MMGL and analyze the empirical results to pave the way for future MMGL research. | Multimodal Graph Learning for Generative Tasks | [
"Minji Yoon",
"Jing Yu Koh",
"Bryan Hooi",
"Russ Salakhutdinov"
] | Workshop/GLFrontiers | 2310.07478 | [
"https://github.com/minjiyoon/mmgl"
] | https://huggingface.co/papers/2310.07478 | 0 | 1 | 0 | 4 | [] | [] | [] | [] | [] | [] | 1 | poster |
null | https://openreview.net/forum?id=XanCwCz4mR | @inproceedings{
kausika2023curiouswalk,
title={CuriousWalk: Enhancing Multi-Hop Reasoning in Graphs with Random Network Distillation},
author={Varun Kausika and Saurabh Jha and Adya Jha and Amy Zhang and Michael Sury},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=XanCwCz4mR}
} | Structured knowledge bases in the forms of graphs often suffer from incompleteness and inaccuracy in representing information. One popular method of densifying graphs involves constructing a reinforcement learning agent that learns to traverse entities and relations in a sequential way from a query entity, according to a query relation until it reaches the desired answer entity. However, these agents are often limited by sparse reward structures of the environment, as well as their inability to find diverse paths from the question to the answer entities. In this paper, we attempt to address these issues by augmenting the agent with intrinsic rewards which can help in exploration as well as offering meaningful feedback at intermediate steps to push the agent in the right direction. | CuriousWalk: Enhancing Multi-Hop Reasoning in Graphs with Random Network Distillation | [
"Varun Kausika",
"Saurabh Jha",
"Adya Jha",
"Amy Zhang",
"Michael Sury"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=VAATRBCyPg | @inproceedings{
chowdhury2023sparse,
title={Sparse but Strong: Crafting Adversarially Robust Graph Lottery Tickets},
author={Subhajit Dutta Chowdhury and Zhiyu Ni and Qingyuan Peng and Souvik Kundu and Pierluigi Nuzzo},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=VAATRBCyPg}
} | Graph Lottery Tickets (GLTs), comprising a sparse adjacency matrix and a sparse graph neural network (GNN), can significantly reduce the inference latency and compute footprint compared to their dense counterparts. Despite these benefits, their performance against adversarial structure perturbations remains to be fully explored. In this work, we first investigate the resilience of GLTs against different structure perturbation attacks and observe that they are highly vulnerable and show a large drop in classification accuracy. Based on this observation, we then present an adversarially robust graph sparsification (ARGS) framework that prunes the adjacency matrix and the GNN weights by optimizing a novel loss function capturing the graph homophily property and information associated with both the true labels of the train nodes and the pseudo labels of the test nodes. By iteratively applying ARGS to prune both the perturbed graph adjacency matrix and the GNN model weights, we can find adversarially robust graph lottery tickets that are highly sparse yet achieve competitive performance under different untargeted training-time structure attacks. Evaluations conducted on various benchmarks, considering different poisoning structure attacks, namely, PGD, MetaAttack, Meta-PGD, and PR-BCD demonstrate that the GLTs generated by ARGS can significantly improve the robustness, even when subjected to high levels of sparsity. | Sparse but Strong: Crafting Adversarially Robust Graph Lottery Tickets | [
"Subhajit Dutta Chowdhury",
"Zhiyu Ni",
"Qingyuan Peng",
"Souvik Kundu",
"Pierluigi Nuzzo"
] | Workshop/GLFrontiers | 2312.06568 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=V5eDwEDfXT | @inproceedings{
koke2023resolvnet,
title={ResolvNet: A Graph Convolutional Network with multi-scale Consistency},
author={Christian Koke and Abhishek Saroha and Yuesong Shen and Marvin Eisenberger and Daniel Cremers},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=V5eDwEDfXT}
} | It is by now a well known fact in the graph learning community that the presence of bottlenecks severely limits the ability of graph neural networks to propagate information over long distances. What so far has not been appreciated is that, counter-intuitively, also the presence of strongly connected sub-graphs may severely restrict information flow in common architectures. Motivated by this observation, we introduce the concept of multi-scale consistency. At the node level this concept refers to the retention of a connected propagation graph even if connectivity varies over a given graph. At the graph-level, multi-scale consistency refers to the fact that distinct graphs describing the same object at different resolutions should be assigned similar feature vectors. As we show, both properties are not satisfied by poular graph neural network architectures. To remedy these shortcomings, we introduce ResolvNet, a flexible graph neural network based on the mathematical concept of resolvents. We rigorously establish its multi-scale consistency theoretically and verify it in extensive experiments on real world data: Here networks based on this ResolvNet architecture prove expressive; out-performing baselines significantly on many tasks; in- and outside the multi-scale setting. | ResolvNet: A Graph Convolutional Network with multi-scale Consistency | [
"Christian Koke",
"Abhishek Saroha",
"Yuesong Shen",
"Marvin Eisenberger",
"Daniel Cremers"
] | Workshop/GLFrontiers | 2310.00431 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=U0EPHUDrot | @inproceedings{
dinh2023on,
title={On the modelling and impact of negative edges in graph convolutional networks for node classification},
author={Thu Trang Dinh and Julia Handl and Luis Ospina-Forero},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=U0EPHUDrot}
} | Signed graphs are important data structures to simultaneously express positive and negative relationships. Their application ranges from structural health monitoring to financial models, where the meaning and properties of negative relationships can play a significant role. In this paper, we provide a comprehensive examination of existing approaches for the integration of signed edges into the Graph Convolutional Network (GCN) framework for node classification. Here we use a combination of theoretical and empirical analysis to gain a deeper understanding of the strengths and limitations of different mechanisms and to identify areas for possible improvement. We compare six different approaches to the integration of negative link information within the framework of the simple GCN. In particular, we analyze sensitivity towards feature noise, negative edge noise and positive edge noise, as well as robustness towards feature scaling and translation, explaining the results obtained on the basis of individual model assumptions and biases. Our findings highlight the importance of capturing the meaning of negative links in a given domain context, and appropriately reflecting it in the choice of GCN model. | On the modelling and impact of negative edges in graph convolutional networks for node classification | [
"Thu Trang Dinh",
"Julia Handl",
"Luis Ospina-Forero"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=TsjGaj45Li | @inproceedings{
sestak2023vnegnn,
title={{VN}-{EGNN}: Equivariant Graph Neural Networks with Virtual Nodes Enhance Protein Binding Site Identification},
author={Florian Sestak and Lisa Schneckenreiter and Sepp Hochreiter and Andreas Mayr and G{\"u}nter Klambauer},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=TsjGaj45Li}
} | Being able to identify regions within or around proteins, to which ligands can
potentially bind, is an essential step to develop new drugs. Binding site iden-
tification methods can now profit from the availability of large amounts of 3D
structures in protein structure databases or from AlphaFold predictions. Current
binding site identification methods rely on geometric deep learning, which takes
geometric invariances and equivariances into account. Such methods turned out
to be very beneficial for physics-related tasks like binding energy or motion tra-
jectory prediction. However, their performance at binding site identification is
still limited, which might be due to limited expressivity or oversquashing effects
of E(n)-Equivariant Graph Neural Networks (EGNNs). Here, we extend EGNNs
by adding virtual nodes and applying an extended message passing scheme. The
virtual nodes in these graphs both improve the predictive performance and can also
learn to represent binding sites. In our experiments, we show that VN-EGNN sets
a new state of the art at binding site identification on three common benchmarks,
COACH420, HOLO4K, and PDBbind2020. | VN-EGNN: Equivariant Graph Neural Networks with Virtual Nodes Enhance Protein Binding Site Identification | [
"Florian Sestak",
"Lisa Schneckenreiter",
"Sepp Hochreiter",
"Andreas Mayr",
"Günter Klambauer"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
||
null | https://openreview.net/forum?id=T40bRpEd6P | @inproceedings{
jiang2023hierarchical,
title={Hierarchical Relationships: A New Perspective to Enhance Scene Graph Generation},
author={Bowen Jiang and Camillo Taylor},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=T40bRpEd6P}
} | This paper presents a finding that leveraging the hierarchical structures among labels for relationships and objects can substantially improve the performance of scene graph generation systems. The focus of this work is to create an informative hierarchical structure that can divide object and relationship categories into disjoint super-categories in a systematic way. Specifically, we introduce a Bayesian prediction head to jointly predict the super-category of relationships between a pair of object instances, as well as the detailed relationship within that super-category simultaneously, facilitating more informative predictions. The resulting model exhibits the capability to produce a more extensive set of predicates beyond the dataset annotations, and to tackle the prevalent issue of low annotation quality. While our paper presents preliminary findings, experiments on the Visual Genome dataset show its strong performance, particularly in predicate classifications and zero-shot settings, that demonstrates the promise of our approach. | Hierarchical Relationships: A New Perspective to Enhance Scene Graph Generation | [
"Bowen Jiang",
"Camillo Taylor"
] | Workshop/GLFrontiers | 2303.06842 | [
"https://github.com/bowen-upenn/scene_graph_commonsense"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=ScNNo7v4t0 | @inproceedings{
chen2023exploring,
title={Exploring the Potential of Large Language Models ({LLM}s) in Learning on Graph},
author={Zhikai Chen and Haitao Mao and Hang Li and Wei Jin and Hongzhi Wen and Xiaochi Wei and Shuaiqiang Wang and Dawei Yin and Wenqi Fan and Hui Liu and Jiliang Tang},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=ScNNo7v4t0}
} | Learning on Graphs has attracted immense attention due to its wide real-world applications. The most popular pipeline for learning on graphs with textual node attributes primarily relies on Graph Neural Networks (GNNs), and utilizes shallow text embedding as initial node representations, which has limitations in general knowledge and profound semantic understanding. In recent years, Large Language Models (LLMs) have been proven to possess extensive common knowledge and powerful semantic comprehension abilities that have revolutionized existing workflows to handle text data. In this paper, we aim to explore the potential of LLMs in graph machine learning, especially the node classification task, and investigate two possible pipelines: LLMs-as-Enhancers and LLMs-as-Predictors. The former leverages LLMs to enhance nodes' text attributes with their massive knowledge and then generate predictions through GNNs. The latter attempts to directly employ LLMs as standalone predictors. We conduct comprehensive and systematical studies on these two pipelines under various settings. From comprehensive empirical results, we make original observations and find new insights that open new possibilities and suggest promising directions to leverage LLMs for learning on graphs. | Exploring the Potential of Large Language Models (LLMs) in Learning on Graph | [
"Zhikai Chen",
"Haitao Mao",
"Hang Li",
"Wei Jin",
"Hongzhi Wen",
"Xiaochi Wei",
"Shuaiqiang Wang",
"Dawei Yin",
"Wenqi Fan",
"Hui Liu",
"Jiliang Tang"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=ROB6TdsmFa | @inproceedings{
gao2023one,
title={One Node Per User: Node-Level Federated Learning for Graph Neural Networks},
author={Zhidong Gao and Yuanxiong Guo and Yanmin Gong},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=ROB6TdsmFa}
} | Graph Neural Networks (GNNs) training often necessitates gathering raw user data on a central server, which raises significant privacy concerns. Federated learning emerges as a solution, enabling collaborative model training without users directly sharing their raw data. However, integrating federated learning with GNNs presents unique challenges, especially when a client represents a graph node and holds merely a single feature vector. In this paper, we propose a novel framework for node-level federated graph learning. Specifically, we decouple the message-passing and feature vector transformation processes of the first GNN layer, allowing them to be executed separately on the user devices and the cloud server. Moreover, we introduce a graph Laplacian term based on the feature vector's latent representation to regulate the user-side model updates. The experiment results on multiple datasets show that our approach achieves better performance compared with baselines. | One Node Per User: Node-Level Federated Learning for Graph Neural Networks | [
"Zhidong Gao",
"Yuanxiong Guo",
"Yanmin Gong"
] | Workshop/GLFrontiers | 2409.19513 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=QgdnOBnh9i | @inproceedings{
giovanni2023how,
title={How does over-squashing affect the power of {GNN}s?},
author={Francesco Di Giovanni and T. Konstantin Rusch and Michael Bronstein and Andreea Deac and Marc Lackenby and Siddhartha Mishra and Petar Veli{\v{c}}kovi{\'c}},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=QgdnOBnh9i}
} | Graph Neural Networks (GNNs) are the state-of-the-art model for machine learning on graph-structured data.
The most popular class of GNNs operate by exchanging information between adjacent nodes, and are known as Message Passing Neural Networks (MPNNs). While understanding the expressive power of MPNNs is a key question, existing results typically consider settings with uninformative node features. In this paper, we provide a rigorous analysis to determine which function classes of node features can be learned by an MPNN of a given capacity. We do so by measuring the level of \emph{pairwise interactions} between nodes that MPNNs allow for. This measure provides a novel quantitative characterization of the so-called over-squashing effect, which is observed to occur when a large volume of messages is aggregated into fixed-size vectors. Using our measure, we prove that, to guarantee sufficient communication between pairs of nodes, the capacity of the MPNN must be large enough, depending on properties of the input graph structure, such as commute times. For many relevant scenarios, our analysis results in impossibility statements in practice, showing that \emph{over-squashing hinders the expressive power of MPNNs}. Our theory also holds for geometric graphs and hence extends to equivariant MPNNs on point clouds. We validate our analysis through extensive controlled experiments and ablation studies. | How does over-squashing affect the power of GNNs? | [
"Francesco Di Giovanni",
"T. Konstantin Rusch",
"Michael Bronstein",
"Andreea Deac",
"Marc Lackenby",
"Siddhartha Mishra",
"Petar Veličković"
] | Workshop/GLFrontiers | 2306.03589 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=QA3Z1cPDrc | @inproceedings{
georgiev2023beyond,
title={Beyond Erdos-Renyi: Generalization in Algorithmic Reasoning on Graphs},
author={Dobrik Georgiev and Pietro Lio and Jakub Bachurski and Junhua Chen and Tunan Shi},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=QA3Z1cPDrc}
} | Neural algorithmic reasoning excels in many graph algorithms, but assessment mainly focuses on the Erdős-Rényi (ER) graph family. This study delves into graph algorithmic models' generalization across diverse distributions. Testing a leading model exposes overreliance on ER graphs for generalization assessment. We further investigate two scenarios: generalisation to every target distribution and single target distributions. Our results suggest that achieving the former is not as trivial and achieving the latter can be aided selecting source distribution via novel Tree Mover's Distance interpretation. | Beyond Erdos-Renyi: Generalization in Algorithmic Reasoning on Graphs | [
"Dobrik Georgiev",
"Pietro Lio",
"Jakub Bachurski",
"Junhua Chen",
"Tunan Shi"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=PlqQ8V5BfC | @inproceedings{
paliotta2023graph,
title={Graph Neural Networks Go Forward-Forward},
author={Daniele Paliotta and Mathieu Alain and B{\'a}lint M{\'a}t{\'e} and Fran{\c{c}}ois Fleuret},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=PlqQ8V5BfC}
} | We present the Graph Forward-Forward (GFF) algorithm, an extension of the Forward-Forward procedure to graphs, able to handle features distributed over a graph's nodes. This allows training graph neural networks with forward passes only, without backpropagation. Our method is agnostic to the message-passing scheme, and provides a more biologically plausible learning scheme than backpropagation, while also carrying computational advantages. With GFF, graph neural networks are trained greedily layer by layer, using both positive and negative samples. We run experiments on 11 standard graph property prediction tasks, showing how GFF provides an effective alternative to backpropagation for training graph neural networks. This shows in particular that this procedure is remarkably efficient in spite of combining the per-layer training with the locality of the processing in a GNN. | Graph Neural Networks Go Forward-Forward | [
"Daniele Paliotta",
"Mathieu Alain",
"Bálint Máté",
"François Fleuret"
] | Workshop/GLFrontiers | 2302.05282 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=PYcp183GBL | @inproceedings{
garcia2023towards,
title={Towards Particle Flow Event Reconstruction at the Future Circular Collider with {GNN}s},
author={Dolores Garcia and Gregor Kr{\v{z}}manc and Philipp Zehetner and Jan Kieseler and Michele Selvaggi},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=PYcp183GBL}
} | Reconstructing particles properties from raw signals measured in particle physics detectors is a challenging task due to the complex shapes of the showers, variety in density and sparsity. Classical particle reconstruction algorithms in current detectors use a multi-step pipeline, but the increase in data complexity of future detectors will reduce their performance. We consider a geometric graph representation due to the sparsity and difference in density of particle showers. We introduce a dataset for particle level reconstruction at the Future Circular Collider and benchmark the performance of state-of-the-art GNN architectures on this dataset. We show that our pipeline performs with high efficiency and response and discuss how this type of data can further drive the development of novel geometric GNN approaches. | Towards Particle Flow Event Reconstruction at the Future Circular Collider with GNNs | [
"Dolores Garcia",
"Gregor Kržmanc",
"Philipp Zehetner",
"Jan Kieseler",
"Michele Selvaggi"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=PPIVoxynCt | @inproceedings{
chen2023edge,
title={{EDGE}++: Improved Training and Sampling of {EDGE}},
author={Xiaohui Chen and Mingyang Wu and Liping Liu},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=PPIVoxynCt}
} | Traditional graph-generative models like the Stochastic-Block Model (SBM) fall short in capturing complex structures inherent in large graphs. Recently developed deep learning models like NetGAN, CELL, and Variational Graph Autoencoders have made progress but face limitations in replicating key graph statistics. Diffusion-based methods such as EDGE have emerged as promising alternatives, however, they present challenges in computational efficiency and generative performance. In this paper, we propose enhancements to the EDGE model to address these issues. Specifically, we introduce a degree-specific noise schedule that optimizes the number of active nodes at each timestep, significantly reducing memory consumption. Additionally, we present an improved sampling scheme that fine-tunes the generative process, allowing for better control over the similarity between the synthesized and the true network. Our experimental results demonstrate that the proposed modifications not only improve the efficiency but also enhance the accuracy of the generated graphs, offering a robust and scalable solution for graph generation tasks. | EDGE++: Improved Training and Sampling of EDGE | [
"Xiaohui Chen",
"Mingyang Wu",
"Liping Liu"
] | Workshop/GLFrontiers | 2310.14441 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=Nrs8BA84br | @inproceedings{
cha2023on,
title={On the Temperature of Bayesian Graph Neural Networks for Conformal Prediction},
author={Seohyeon Cha and Honggu Kang and Joonhyuk Kang},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=Nrs8BA84br}
} | Accurate uncertainty quantification in graph neural networks (GNNs) is essential, especially in high-stakes domains where GNNs are frequently employed. Conformal prediction (CP) offers a promising framework for quantifying uncertainty by providing $\\textit{valid}$ prediction sets for any black-box model. CP ensures formal probabilistic guarantees that a prediction set contains a true label with a desired probability. However, the size of prediction sets, known as $\\textit{inefficiency}$, is influenced by the underlying model and data generating process. On the other hand, Bayesian learning also provides a credible region based on the estimated posterior distribution, but this region is $\\textit{well-calibrated}$ only when the model is correctly specified. Building on a recent work that introduced a scaling parameter for constructing valid credible regions from posterior estimate, our study explores the advantages of incorporating a temperature parameter into Bayesian GNNs within CP framework. We empirically demonstrate the existence of temperatures that result in more efficient prediction sets. Furthermore, we conduct an analysis to identify the factors contributing to inefficiency and offer valuable insights into the relationship between CP performance and model calibration. | On the Temperature of Bayesian Graph Neural Networks for Conformal Prediction | [
"Seohyeon Cha",
"Honggu Kang",
"Joonhyuk Kang"
] | Workshop/GLFrontiers | 2310.11479 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=NE8Da26bEW | @inproceedings{
xuanyuan2023shedding,
title={Shedding Light on Random Dropping and Oversmoothing},
author={Han Xuanyuan and Tianxiang Zhao and Dongsheng Luo},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=NE8Da26bEW}
} | Graph Neural Networks (GNNs) are widespread in graph representation learning. *Random dropping* approaches, notably DropEdge and DropMessage, claim to alleviate the key issues of overfitting and oversmoothing by randomly removing elements of the graph representation. However, their effectiveness is largely unverified. In this work, we show empirically that they have a limited effect in reducing oversmoothing at test time due to their training time exclusive nature. We show that DropEdge in particular can be seen as a form of training data augmentation, and its benefits to model generalization are not strictly related to oversmoothing, suggesting that in practice, the precise link between oversmoothing and performance is more nuanced than previously thought. We address the limitations of current dropping methods by *learning* to drop via optimizing an information bottleneck, which enables dropping to be performed effectively at test time. | Shedding Light on Random Dropping and Oversmoothing | [
"Han Xuanyuan",
"Tianxiang Zhao",
"Dongsheng Luo"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=LzMWMJlxHg | @inproceedings{
galkin2023towards,
title={Towards Foundation Models for Knowledge Graph Reasoning},
author={Mikhail Galkin and Xinyu Yuan and Hesham Mostafa and Jian Tang and Zhaocheng Zhu},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=LzMWMJlxHg}
} | Foundation models in language and vision have the ability to run inference on any textual and visual inputs thanks to the transferable representations such as a vocabulary of tokens in language. Knowledge graphs (KGs) have different entity and relation vocabularies that generally do not overlap. The key challenge of designing foundation models on KGs is to learn such transferable representations that enable inference on any graph with arbitrary entity and relation vocabularies. In this work, we make a step towards such foundation models and present ULTRA, an approach for learning universal and transferable graph representations. ULTRA builds relational representations as a function conditioned on their interactions. Such a conditioning strategy allows a pre-trained ULTRA model to inductively generalize to any unseen KG with any relation vocabulary and to be fine-tuned on any graph. Conducting link prediction experiments on 57 different KGs, we find that the zero-shot inductive inference performance of a single pre-trained ULTRA model on unseen graphs of various sizes is often on par or better than strong baselines trained on specific graphs. Fine-tuning further boosts the performance. | Towards Foundation Models for Knowledge Graph Reasoning | [
"Mikhail Galkin",
"Xinyu Yuan",
"Hesham Mostafa",
"Jian Tang",
"Zhaocheng Zhu"
] | Workshop/GLFrontiers | 2310.04562 | [
"https://github.com/DeepGraphLearning/ULTRA"
] | https://huggingface.co/papers/2310.04562 | 1 | 3 | 1 | 5 | [
"mgalkin/ultra_50g",
"mgalkin/ultra_3g",
"mgalkin/ultra_4g"
] | [] | [] | [
"mgalkin/ultra_50g",
"mgalkin/ultra_3g",
"mgalkin/ultra_4g"
] | [] | [] | 1 | poster |
null | https://openreview.net/forum?id=KIISfvU806 | @inproceedings{
zhou2023a,
title={A Multi-Task Perspective for Link Prediction with New Relation Types and Nodes},
author={Jincheng Zhou and Beatrice Bevilacqua and Bruno Ribeiro},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=KIISfvU806}
} | The task of inductive link prediction in (discrete) attributed multigraphs infers missing attributed links (relations) between nodes in new test multigraphs. Traditional relational learning methods face the challenge of limited generalization to test multigraphs containing both novel nodes and novel relation types not seen in training. Recently, under the only assumption that all relation types share the same structural predictive patterns (single task), Gao et al. (2023) proposed a link prediction method using the theoretical concept of double equivariance (equivariance for nodes & relation types), in contrast to the (single) equivariance (only for nodes) used to design Graph Neural Networks (GNNs). In this work we further extend the double equivariance concept to multi-task double equivariance, where we define link prediction in attributed multigraphs that can have distinct and potentially conflicting predictive patterns for different sets of relation types (multiple tasks). Our empirical results on real-world datasets demonstrate that our approach can effectively generalize to test graphs with multi-task structures without access to additional information. | A Multi-Task Perspective for Link Prediction with New Relation Types and Nodes | [
"Jincheng Zhou",
"Beatrice Bevilacqua",
"Bruno Ribeiro"
] | Workshop/GLFrontiers | 2307.06046 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=JTRErQb2oN | @inproceedings{
ouyang2023prompt,
title={Prompt Learning Unlocked for App Promotion in the Wild},
author={Zhongyu Ouyang and Shifu Hou and Shang Ma and Chaoran Chen and Chunhui Zhang and Toby Li and Xusheng Xiao and Chuxu Zhang and Yanfang Ye},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=JTRErQb2oN}
} | In recent times, mobile apps have increasingly incorporated app promotion ads to promote other apps, raising cybersecurity and online commerce concerns related to societal trust and recommendation systems. To effectively discover the intricate nature of the app promotion graph data, we center around the graph completion task, aiming to learn the connection patterns among diverse relations and entities. However, accurately deciphering the connection patterns in such a large and diverse graph presents significant challenges for deep learning models. To overcome these challenges, we introduce Prompt Promotion, a transformer-based framework that unlocks prompt learning capabilities by incorporating metapath- and embedding-based prompts that provide valuable hints to guide the model's predictions for undetermined connection patterns. Experimental results show that our Prompt Promotion model represents a pioneering prompt-based capability in effectively completing the app promotion graph. It not only demonstrates superior performance in heterogeneous graph completion in real-world scenarios, but also exhibits strong generalization capabilities for diverse, complex, and noisy connection patterns when paired with their respective prompts. | Prompt Learning Unlocked for App Promotion in the Wild | [
"Zhongyu Ouyang",
"Shifu Hou",
"Shang Ma",
"Chaoran Chen",
"Chunhui Zhang",
"Toby Li",
"Xusheng Xiao",
"Chuxu Zhang",
"Yanfang Ye"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=It9QWbVe02 | @inproceedings{
chen2023uncertaintyaware,
title={Uncertainty-Aware Robust Learning on Noisy Graphs},
author={Shuyi Chen and Kaize Ding and Shixiang Zhu},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=It9QWbVe02}
} | Graph neural networks have shown impressive capabilities in solving various graph learning tasks, particularly excelling in node classification. However, their effectiveness can be hindered by the challenges arising from the widespread existence of noisy measurements associated with the topological or nodal information present in real-world graphs. These inaccuracies in observations can corrupt the crucial patterns within the graph data, ultimately resulting in undesirable performance in practical applications. To address these issues, this paper proposes a novel uncertainty-aware graph learning framework motivated by distributionally robust optimization. The framework aims to alleviate the challenges by considering the distributional uncertainty associated with the graph data. Specifically, we use a graph neural network-based encoder to embed the node features and find the optimal node embeddings by minimizing the worst-case risk through a minimax formulation. Such an uncertainty-aware learning process leads to improved node representations and a more robust graph predictive model that effectively mitigates the impact of uncertainty arising from data noise. The learned LFDs also provide a means to quantify the predictive uncertainty, which is valuable in some uncertainty-sensitive scenarios where incorrect decisions can have severe consequence. In addition, we adopt the idea of differentiable optimization and develop an end-to-end learning algorithm that seamlessly integrates graph learning and distributionally robust optimization. Our experimental result shows that the proposed framework achieves superior predictive performance compared to the state-of-the-art baselines under various noisy settings. | Uncertainty-Aware Robust Learning on Noisy Graphs | [
"Shuyi Chen",
"Kaize Ding",
"Shixiang Zhu"
] | Workshop/GLFrontiers | 2306.08210 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=I5hf3opvgK | @inproceedings{
roy2023gadebm,
title={{GAD}-{EBM}: Graph Anomaly Detection using Energy-Based Models},
author={Amit Roy and Juan Shu and Olivier Elshocht and Jeroen Smeets and Ruqi Zhang and Pan Li},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=I5hf3opvgK}
} | Graph Anomaly Detection (GAD) is essential in fields ranging from network security, bioinformatics to finance. Previous works often adopt auto-encoders to compute reconstruction errors for anomaly detection: anomalies are hard to be reconstructed. In this work, we revisit the first principle for anomaly detection, i.e., the Neyman-Pearson rule, where the optimal anomaly detector is based on the likelihood of a data point given the normal distribution of data. However, in practice, the distribution is often unknown and the estimation of the distribution of graph-structured data may be hard. Moreover, the likelihood computation of a graph-structured data point may be challenging as well. In this paper, we propose a novel approach GAD-EBM that can estimate the distribution of graphs and compute likelihoods efficiently by using Energy-Based Models (EBMs) over graphs. GAD-EBM approaches the likelihood of a rooted subgraph of node v, and further can leverage the likelihood to accurately identify whether node v is anomalous or not. Traditional score matching for training EBMs may not be used to apply EBMs that model the distribution of graphs because of complicated discreteness and multi-modality of graph data. We propose a Subgraph Score Matching (SSM) approach, which is specifically designed for graph data based on a novel framework of neighborhood state-space graphs. Experimentation conducted on six real-world datasets validates the effectiveness and efficiency of GAD-EBM and the source code for GAD-EBM is openly available. | GAD-EBM: Graph Anomaly Detection using Energy-Based Models | [
"Amit Roy",
"Juan Shu",
"Olivier Elshocht",
"Jeroen Smeets",
"Ruqi Zhang",
"Pan Li"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=HKdsrm5nCW | @inproceedings{
kataria2023linear,
title={Linear Complexity Framework for Feature-Aware Graph Coarsening via Hashing},
author={Mohit Kataria and Aditi Khandelwal and Rocktim Das and Sandeep Kumar and Jayadeva Jayadeva},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=HKdsrm5nCW}
} | Large-scale graphs are increasingly common in various applications, leading to significant computational challenges in data processing and analysis. To address this, coarsening algorithms are employed to reduce graph size while preserving key properties. However, existing methods for large-scale graphs are computationally intensive, undermining the coarsening goal. Additionally, real-world graphs often contain node-specific features or contexts, which current coarsening approaches overlook, focusing solely on structural information like adjacency matrices. This limitation may not suit downstream tasks reliant on node features. In this paper, we introduce a Feature-Aware graph Coarsening algorithm via Hashing, called FACH, inspired by locality sensitive hashing to coarsen the graph based on the node features. To our knowledge, this is the first-ever method that coarsens a graph with node features in linear time. FACH is over 7× faster than the quickest and around 150× faster than the existing techniques for datasets like Coauthor Physics which has 34,493 nodes. We also demonstrate the efficacy of the proposed framework in terms of superior run-time complexity. The coarsened graph obtained by our method also preserves the spectral properties of the original graph while achieving massive improvement in time-complexity of coarsening which is the primary goal of our study. We showcase the effectiveness of FACH for the downstream task by evaluating the performance on scalable training of graph neural networks using coarsened data on benchmark real-world datasets. | Linear Complexity Framework for Feature-Aware Graph Coarsening via Hashing | [
"Mohit Kataria",
"Aditi Khandelwal",
"Rocktim Das",
"Sandeep Kumar",
"Jayadeva Jayadeva"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=EVp40Cz0PR | @inproceedings{
g2023satg,
title={{SATG} : Structure Aware Transformers on Graphs for Node Classification},
author={Sumedh B G and Sanjay Patnala and Himil Vasava and Akshay Sethi and Sonia Gupta},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=EVp40Cz0PR}
} | Transformers have achieved state-of-the-art performance in the fields of Computer Vision (CV) and Natural Language Processing (NLP). Inspired by this, architectures have come up in recent times that incorporate transformers into the domain of graph neural networks. Most of the existing Graph Transformers either take a set of all the nodes as an input sequence leading to quadratic time complexity or they take only one hop or k-hop neighbours as the input sequence, thereby completely ignoring any long-range interactions. To this end, we propose Structure Aware Transformer on Graphs (SATG), where we capture both short-range and long-range interactions in a computationally efficient manner. When it comes to dealing with non-euclidean spaces like graphs, positional encoding becomes an integral component to provide structural knowledge to the transformer. Upon observing the shortcomings of the existing set of positional encodings, we introduce a new class of positional encodings trained on a Neighbourhood Contrastive Loss that effectively captures the entire topology of the graph. We also introduce a method to effectively capture long-range interactions without having a quadratic time complexity. Extensive experiments done on five benchmark datasets show that SATG consistently outperforms GNNs by a substantial margin and also successfully outperforms other Graph Transformers. | SATG : Structure Aware Transformers on Graphs for Node Classification | [
"Sumedh B G",
"Sanjay Patnala",
"Himil Vasava",
"Akshay Sethi",
"Sonia Gupta"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=BaxFC3z9R6 | @inproceedings{
li2023what,
title={What Improves the Generalization of Graph Transformer? A Theoretical Dive into Self-attention and Positional Encoding},
author={Hongkang Li and Meng Wang and Tengfei Ma and Sijia Liu and ZAIXI ZHANG and Pin-Yu Chen},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=BaxFC3z9R6}
} | Graph Transformers, which incorporate self-attention and positional encoding, have recently emerged as a powerful architecture for various graph learning tasks. Despite their impressive performance, the complex non-convex interactions across layers and the recursive graph structure have made it challenging to establish a theoretical foundation for learning and generalization. This study introduces the first theoretical investigation of a shallow Graph Transformer for semi-supervised node classification, comprising a self-attention layer with relative positional encoding and a two-layer perception. Focusing on a graph data model with discriminative nodes that determine node labels and non-discriminative nodes that are class-irrelevant, we characterize the sample complexity required to achieve a zero generalization error by training with stochastic gradient descent (SGD). This paper provides the quantitative characterization of the sample complexity and number of iterations for convergence dependent on the fraction of discriminative nodes, the dominant patterns, the fraction of erroneous labels, and the initial model errors. Furthermore, we demonstrate that self-attention and positional encoding enhance generalization by making the attention map sparse and promoting the core neighborhood during training, which explains the superior feature representation of Graph Transformers. Our theoretical results are supported by empirical experiments on synthetic and real-world benchmarks. | What Improves the Generalization of Graph Transformer? A Theoretical Dive into Self-attention and Positional Encoding | [
"Hongkang Li",
"Meng Wang",
"Tengfei Ma",
"Sijia Liu",
"ZAIXI ZHANG",
"Pin-Yu Chen"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=9bD3oqcD6g | @inproceedings{
sohn2023todflow,
title={{TOD}-Flow: Modeling the Structure of Task-Oriented Dialogues},
author={Sungryull Sohn and Yiwei Lyu and Anthony Liu and Lajanugen Logeswaran and Dong-Ki Kim and Dongsub Shim and Honglak Lee},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=9bD3oqcD6g}
} | Task-Oriented Dialogue (TOD) systems have become crucial components in interactive artificial intelligence applications. While recent advances have capitalized on pre-trained language models (PLMs), they exhibit limitations regarding transparency and controllability. To address these challenges, we propose a novel approach focusing on inferring the TOD-flow graph from dialogue data annotated with dialog acts, uncovering the underlying task structure in the form of a graph. The inferred TOD-flow graph can be easily integrated with any dialogue model to improve its prediction performance, transparency, and controllability. Our TOD-flow graph learns what a model can, should, and should not predict, effectively reducing the search space and providing a rationale for the model's prediction.
We show that the proposed TOD-flow graph better resemble human-annotated graphs compared to prior approaches.
Furthermore, when combined with several dialogue policies and end-to-end dialogue models, we demonstrate that our approach significantly improves dialog act classification and end-to-end response generation performance in the MultiWOZ and SGD benchmarks. | TOD-Flow: Modeling the Structure of Task-Oriented Dialogues | [
"Sungryull Sohn",
"Yiwei Lyu",
"Anthony Liu",
"Lajanugen Logeswaran",
"Dong-Ki Kim",
"Dongsub Shim",
"Honglak Lee"
] | Workshop/GLFrontiers | 2312.04668 | [
"https://github.com/srsohn/tod-flow"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=9ADkymyCPA | @inproceedings{
subramonian2023networked,
title={Networked Inequality: Preferential Attachment Bias in Graph Neural Network Link Prediction},
author={Arjun Subramonian and Levent Sagun and Yizhou Sun},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=9ADkymyCPA}
} | Graph neural network (GNN) link prediction is increasingly deployed in citation, collaboration, and online social networks to recommend academic literature, collaborators, and friends. While prior research has investigated the dyadic fairness of GNN link prediction, the within-group fairness and ``rich get richer'' dynamics of link prediction remain underexplored. However, these aspects have significant consequences for degree and power imbalances in networks. In this paper, we shed light on how degree bias in networks affects Graph Convolutional Network (GCN) link prediction. In particular, we theoretically uncover that GCNs with a symmetric normalized graph filter have a within-group preferential attachment bias. We validate our theoretical analysis on real-world citation, collaboration, and online social networks. We further bridge GCN's preferential attachment bias with unfairness in link prediction and propose a new within-group fairness metric. This metric quantifies disparities in link prediction scores between social groups, towards combating the amplification of degree and power disparities. Finally, we propose a simple training-time strategy to alleviate within-group unfairness, and we show that it is effective on citation, online social, and credit networks. | Networked Inequality: Preferential Attachment Bias in Graph Neural Network Link Prediction | [
"Arjun Subramonian",
"Levent Sagun",
"Yizhou Sun"
] | Workshop/GLFrontiers | 2309.17417 | [
"https://github.com/arjunsubramonian/link_bias_amplification"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=947KhgKKGG | @inproceedings{
wang2023towards,
title={Towards Flexible, Efficient, and Effective Tensor Product Networks},
author={Nanxiang Wang and Chen Lin and Michael Bronstein and Philip Torr},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=947KhgKKGG}
} | Geometric graph neural networks have showcased exceptional performance in modelling geometric data. These models rely heavily on equivariant operations, encompassing vital techniques such as scalarization and the Clebsch-Gordan tensor product. However, tensor-product-based architectures face substantial computational challenges as the representation order increases, significantly limiting their versatility. Moreover, the interpretability of interactions between steerable components remains elusive. In contrast, scalarization methods benefit from cost-efficient invariant scalar operations while still being capable of outperforming certain tensor-product-based models. To bridge the gap between these approaches, we introduce a conceptual framework that emphasizes the potential flexibility in designing tensor product networks. To provide guidance for efficient framework design and gain deeper insights into steerable components, we conduct a preliminary investigation by pruning tensor product interactions. This approach enables us to directly assess the redundancy and significance of steerable components, paving the way for efficient and effective designs. | Towards Flexible, Efficient, and Effective Tensor Product Networks | [
"Nanxiang Wang",
"Chen Lin",
"Michael Bronstein",
"Philip Torr"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=7xlWBzaeJN | @inproceedings{
li2023chatpathway,
title={ChatPathway: Conversational Large Language Models for Biology Pathway Detection},
author={Yanjing Li and Hannan Xu and Haiteng Zhao and Hongyu Guo and Shengchao Liu},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=7xlWBzaeJN}
} | Biological pathways, like protein-protein interactions and metabolic networks, are vital for understanding diseases and drug development. Some databases such as KEGG are designed to store and map these pathways. However, many bioinformatics methods face limitations due to database constraints, and certain deep learning models struggle with the complexities of biochemical reactions involving large molecules and diverse enzymes. Importantly, the thorough exploration of biological pathways demands a deep understanding of scientific literature and past research. Despite this, recent advancements in Large Language Models (LLMs), especially ChatGPT, show promise. We first restructured data from KEGG and augmented it with molecule structural and functional information sourced from UniProt and PubChem. Our study evaluated LLMs, particularly GPT-3.5-turbo and Galactica, in predicting biochemical reactions and pathways using our constructed data. We also assessed its ability to predict novel pathways, not covered in its training dataset, using findings from recently published studies. While GPT demonstrated strengths in pathway mapping, Galactica encountered challenges. This research emphasizes the potential of merging LLMs with biology, suggesting a harmonious blend of human expertise and AI in decoding biological systems. | ChatPathway: Conversational Large Language Models for Biology Pathway Detection | [
"Yanjing Li",
"Hannan Xu",
"Haiteng Zhao",
"Hongyu Guo",
"Shengchao Liu"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
||
null | https://openreview.net/forum?id=7CAJpRo1Q8 | @inproceedings{
fatemi2023talk,
title={Talk like a Graph: Encoding Graphs for Large Language Models},
author={Bahare Fatemi and Jonathan Halcrow and Bryan Perozzi},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=7CAJpRo1Q8}
} | Graphs are a powerful tool for representing and analyzing complex relationships in real-world applications such as social networks, recommender systems, and computational finance. Reasoning on graphs is essential for drawing inferences about the relationships between entities in a complex system, and to identify hidden patterns and trends. Despite the remarkable progress in automated reasoning with natural text, reasoning on graphs with large language models (LLMs) remains an understudied problem. In this work, we perform the first comprehensive study of encoding graph-structured data as text for consumption by LLMs. We show that LLM performance on graph reasoning tasks varies on three fundamental levels: (1) the graph encoding method, (2) the nature of the graph task itself, and (3) interestingly, the very structure of the graph considered. These novel results provide valuable insight on strategies for encoding graphs as text. Using these insights we illustrate how the correct choice of encoders can boost performance on graph reasoning tasks inside LLMs by 4.8% to 61.8%, depending on the task. | Talk like a Graph: Encoding Graphs for Large Language Models | [
"Bahare Fatemi",
"Jonathan Halcrow",
"Bryan Perozzi"
] | Workshop/GLFrontiers | 2310.04560 | [
""
] | https://huggingface.co/papers/2310.04560 | 0 | 3 | 0 | 3 | [] | [] | [] | [] | [] | [] | 1 | poster |
null | https://openreview.net/forum?id=6UuNTxAB1t | @inproceedings{
inae2023motifaware,
title={Motif-aware Attribute Masking for Molecular Graph Pre-training},
author={Eric Inae and Gang Liu and Meng Jiang},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=6UuNTxAB1t}
} | Attribute reconstruction is used to predict node or edge features in the pre-training of graph neural networks. Given a large number of molecules, they learn to capture structural knowledge, which is transferable for various downstream property prediction tasks and vital in chemistry, biomedicine, and material science. Previous strategies that randomly select nodes to do attribute masking leverage the information of local neighbors. However, the over-reliance of these neighbors inhibits the model's ability to learn long-range dependencies from higher-level substructures. For example, the model would learn little from predicting three carbon atoms in a benzene ring based on the other three but could learn more from the inter-connections between the functional groups, or called chemical motifs. In this work, we propose and investigate motif-aware attribute masking strategies to capture long-range inter-motif structures by leveraging the information of atoms in neighboring motifs. Once each graph is decomposed into disjoint motifs, the features for every node within a sample motif are masked. The graph decoder then predicts the masked features of each node within the motif for reconstruction. We evaluate our approach on eight molecular property prediction datasets and demonstrate its advantages. | Motif-aware Attribute Masking for Molecular Graph Pre-training | [
"Eric Inae",
"Gang Liu",
"Meng Jiang"
] | Workshop/GLFrontiers | 2309.04589 | [
"https://github.com/einae-nd/moama-dev"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=60fJkjHV0r | @inproceedings{
abbahaddou2023graph,
title={Graph Neural Networks on Discriminative Graphs of Words},
author={Yassine ABBAHADDOU and Johannes Lutzeyer and Michalis Vazirgiannis},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=60fJkjHV0r}
} | In light of the recent success of Graph Neural Networks (GNNs) and their ability to perform inference on complex data structures, many studies apply GNNs to the task of text classification. In most previous methods, a heterogeneous graph, containing both word and document nodes, is constructed using the entire corpus and a GNN is used to classify document nodes. In this work, we explore a new Discriminative Graph of Words Graph Neural Network (DGoW-GNN) approach encapsulating both a novel discriminative graph construction and model to classify text. In our graph construction, containing only word nodes and no document nodes, we split the training corpus into disconnected subgraphs according to their labels and weight edges by the pointwise mutual information of the represented words. Our graph construction, for which we provide theoretical motivation, allows us to reformulate the task of text classification as the task of walk classification. We also propose a new model for the graph-based classification of text, which combines a GNN and a sequence model. We evaluate our approach on seven benchmark datasets and find that it is outperformed by several state-of-the-art baseline models. We analyse reasons for this performance difference and hypothesise under which conditions it is likely to change. Our code is publicly available at: https://github.com/abbahaddou/DGOW. | Graph Neural Networks on Discriminative Graphs of Words | [
"Yassine ABBAHADDOU",
"Johannes Lutzeyer",
"Michalis Vazirgiannis"
] | Workshop/GLFrontiers | 2410.20469 | [
"https://github.com/abbahaddou/DGOW"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=5RA0ysv8PX | @inproceedings{
trivedi2023estimating,
title={Estimating Epistemic Uncertainty of Graph Neural Networks using Stochastic Centering},
author={Puja Trivedi and Mark Heimann and Rushil Anirudh and Danai Koutra and Jayaraman J. Thiagarajan},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=5RA0ysv8PX}
} | While graph neural networks (GNNs) are widely used for node and graph representation learning tasks, the reliability of GNN uncertainty estimates under distribution shifts remains relatively under-explored. Indeed, while \textit{post-hoc} calibration strategies can be used to improve in-distribution calibration, they need not also improve calibration under distribution shift. However, techniques which produce GNNs with better \textit{intrinsic} uncertainty estimates are particularly valuable, as they can always be combined with post-hoc strategies later. Therefore, in this work, we propose G-$\Delta$UQ, a novel training framework designed to improve intrinsic GNN uncertainty estimates. Our framework adapts the principle of stochastic data centering to graph data through novel graph anchoring strategies, and is able to support partially stochastic GNNs. While, the prevalent wisdom is that fully stochastic networks are necessary to obtain reliable estimates, we find that the functional diversity induced by our anchoring strategies when sampling hypotheses renders this unnecessary and allows us to support \gduq~ on pretrained models. Indeed, through extensive evaluation under covariate, concept and graph size shifts, we show that G-$\Delta$UQ leads to better calibrated GNNs for node and graph classification. Further, it also improves performance on the uncertainty-based tasks of out-of-distribution detection and generalization gap estimation. Overall, our work provides insights into uncertainty estimation for GNNs, and demonstrates the utility of G-$\Delta$UQ in obtaining reliable estimates. | Estimating Epistemic Uncertainty of Graph Neural Networks using Stochastic Centering | [
"Puja Trivedi",
"Mark Heimann",
"Rushil Anirudh",
"Danai Koutra",
"Jayaraman J. Thiagarajan"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=4tGqks76l7 | @inproceedings{
hajiramezanali2023on,
title={On the Consistency of {GNN} Explainability Methods},
author={Ehsan Hajiramezanali and Sepideh Maleki and Alex Tseng and Aicha BenTaieb and Gabriele Scalia and Tommaso Biancalani},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=4tGqks76l7}
} | Despite the widespread utilization of post-hoc explanation methods for graph neural networks (GNNs) in high-stakes settings, there has been a lack of comprehensive evaluation regarding their quality and reliability. This evaluation is challenging primarily due to non-Euclidean nature of the data, arbitrary size, and complex topological structure. In this context, we argue that the \emph{consistency} of GNN explanations, denoting the ability to produce similar explanations for input graphs with minor structural changes that do not alter their output predictions, is a key requirement for effective post-hoc GNN explanations. To fulfill this gap, we introduce a novel metric based on Fused Gromov-Wasserstein distance to quantify consistency. | On the Consistency of GNN Explainability Methods | [
"Ehsan Hajiramezanali",
"Sepideh Maleki",
"Alex Tseng",
"Aicha BenTaieb",
"Gabriele Scalia",
"Tommaso Biancalani"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=421Y6z81UZ | @inproceedings{
vora2023gnn,
title={{GNN} Predictions on k-hop Egonets Boosts Adversarial Robustness},
author={Jian Vora},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=421Y6z81UZ}
} | Like many other deep learning models, Graph Neural Networks (GNNs) have
been shown to be susceptible to adversarial attacks, i.e., the addition of crafted
imperceptible noise to input data changes the model predictions drastically. We
propose a very simple method k-HOP-PURIFY which makes node predictions on a
k-hop Egonet centered at the node instead of the entire graph boosts adversarial accuracies. This could be used both as i) a post-processing step after applying popular
defenses or ii) as a standalone defense method which is comparable to many other
competitors. The method is extremely lightweight and scalable (takes 4 lines of
code to implement) unlike many other defense methods which are computationally
expensive or rely on heuristics. We show performance gains through extensive experimentation across various types of attacks (poison/evasion, targetted/untargeted),
perturbation rates, and defenses implemented in the DeepRobust Library. | GNN Predictions on k-hop Egonets Boosts Adversarial Robustness | [
"Jian Vora"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=1JkgXzKKdo | @inproceedings{
younesian2023grapes,
title={{GRAPES}: Learning to Sample Graphs for Scalable Graph Neural Networks},
author={Taraneh Younesian and Thiviyan Thanapalasingam and Emile van Krieken and Daniel Daza and Peter Bloem},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=1JkgXzKKdo}
} | Graph neural networks (GNNs) learn the representation of nodes in a graph by aggregating the neighborhood information in various ways. As these networks grow in depth, their receptive field grows exponentially due to the increase in neighborhood sizes, resulting in high memory costs. Graph sampling solves memory issues in GNNs by sampling a small ratio of the nodes in the graph. This way, GNNs can scale to much larger graphs. Most sampling methods focus on fixed sampling heuristics, which may not generalize to different structures or tasks. We introduce GRAPES, an adaptive graph sampling method that learns to identify sets of influential nodes for training a GNN classifier. GRAPES uses a GFlowNet to learn node sampling probabilities given the classification objectives. We evaluate GRAPES across several small- and large-scale graph benchmarks and demonstrate its effectiveness in accuracy and scalability. In contrast to existing sampling methods, GRAPES maintains high accuracy even with small sample sizes and, therefore, can scale to very large graphs. Our code is publicly available at https://anonymous.4open.science/r/GRAPES. | GRAPES: Learning to Sample Graphs for Scalable Graph Neural Networks | [
"Taraneh Younesian",
"Thiviyan Thanapalasingam",
"Emile van Krieken",
"Daniel Daza",
"Peter Bloem"
] | Workshop/GLFrontiers | 2310.03399 | [
"https://github.com/dfdazac/grapes"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=0oY0LPPH4e | @inproceedings{
zhang2023robust,
title={Robust Hierarchical Scene Graph Generation},
author={Ce Zhang and Simon Stepputtis and Joseph Campbell and Katia Sycara and Yaqi Xie},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=0oY0LPPH4e}
} | The ability to quickly understand scenes from visual observations via structured representations, known as Scene Graph Generation (SGG), is a crucial component of perception models. Despite recent advancements, most existing models assume perfect observations, an often-unrealistic condition in real-world scenarios. Such models can struggle with visual inputs affected by natural corruptions such as sunlight glare, extreme weather conditions, and smoke.
Drawing inspiration from human hierarchical reasoning skills (i.e., from higher to lower levels) as a defense against corruption, we propose a new framework called Hierarchical Knowledge Enhanced Robust Scene Graph Generation (HiKER-SGG). First, we create a hierarchical knowledge graph, facilitating machine comprehension of this structured knowledge. Then we bridge between the constructed graph and the initial scene graph and perform message passing for hierarchical graph reasoning. Finally, we propose a hierarchical prediction head to enable the model to predict from a higher to lower level, thus enhancing robustness against corruptions that frequently impact only fine-grained details. Experiments on various settings confirm the superior performance of the proposed framework with both clean and corrupted images. | Robust Hierarchical Scene Graph Generation | [
"Ce Zhang",
"Simon Stepputtis",
"Joseph Campbell",
"Katia Sycara",
"Yaqi Xie"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=0lZjoxTZFb | @inproceedings{
christie2023higherorder,
title={Higher-Order Expander Graph Propagation},
author={Thomas Christie and Yu He},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=0lZjoxTZFb}
} | Graph neural networks operate on graph-structured data via exchanging messages along edges. One limitation of this message passing paradigm is the over-squashing problem. Over-squashing occurs when messages from a node's expanded receptive field are compressed into fixed-size vectors, potentially causing information loss. To address this issue, recent works have explored using expander graphs, which are highly-connected sparse graphs with low diameters, to perform message passing. However, current methods on expander graph propagation only consider pair-wise interactions, ignoring higher-order structures in complex data. To explore the benefits of capturing these higher-order correlations while still leveraging expander graphs, we introduce higher-order expander graph propagation. We propose two methods for constructing bipartite expanders and evaluate their performance on both synthetic and real-world datasets. | Higher-Order Expander Graph Propagation | [
"Thomas Christie",
"Yu He"
] | Workshop/GLFrontiers | 2311.07966 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=0YeJyvv2rO | @inproceedings{
yue2023mcgc,
title={{MCGC}: an {MLP}-based supervised Contrastive learning framework for Graph Classification},
author={Xiao Yue and Bo Liu and Andrew Meng and Guangzhi Qu},
booktitle={NeurIPS 2023 Workshop: New Frontiers in Graph Learning},
year={2023},
url={https://openreview.net/forum?id=0YeJyvv2rO}
} | Graph Neural Networks (GNNs) have been widely used for tasks involving graph-structured data. These networks create matrix representations of graphs by aggregating nodes information from neighbor nodes recursively. Integrating with contrastive learning, graph contrastive learning has shown enhanced performance on graph-level tasks. However, architectures of graph contrastive learning frameworks become complicated due to the sophisticated structures of GNN-based encoder and necessity of both encoder and projection head. In this paper, we proposed a significantly simplified MLP-based supervised contrastive learning framework for graph classification tasks, coined as MCGC, which does not incorporate any GNN layers. Experimental results on graph benchmark datasets and ablation studies indicate that, despite not utilizing GNN layers, our framework achieved comparable or even superior performance on graph classification tasks against some state-of-the-art models. | MCGC: an MLP-based supervised Contrastive learning framework for Graph Classification | [
"Xiao Yue",
"Bo Liu",
"Andrew Meng",
"Guangzhi Qu"
] | Workshop/GLFrontiers | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=zsNbN3NkzR | @inproceedings{
anonymous2023irreducible,
title={Irreducible Curriculum for Language Model Pretraining},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=zsNbN3NkzR}
} | Automatic data selection and curriculum design for training large language models is challenging, with only a few existing methods showing improvements over standard training.
Furthermore, current schemes focus on domain-level selection, overlooking the more fine-grained contributions of each individual training point. It is difficult to apply traditional datapoint selection methods on large language models: most online batch selection methods perform two-times forward or backward passes, which introduces considerable extra costs with large-scale models.
To mitigate these obstacles, we propose irreducible curriculum as a curriculum learning algorithm for language model pretraining, which prioritizes samples with higher learnability. Specifically, to avoid prohibitive extra computation overhead, we simulate the sample loss along the main model's training trajectory using a small-scale proxy model. Our experiments on the RedPajama-1B dataset demonstrate a consistent improvement on validation perplexity across all 7 domains compared to random uniform baseline and the anti-curriculum strategy. Our method also reduces the sharpness of the network and illustrates a better 5-shot accuracy on MMLU benchmarks. | Irreducible Curriculum for Language Model Pretraining | null | Workshop/ATTRIB | 2310.15389 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=xjmOQRf1OI | @inproceedings{
anonymous2023evaluating,
title={Evaluating the Utility of Model Explanations for Model Development},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=xjmOQRf1OI}
} | One of the motivations for explainable AI is to allow humans to make better and more informed decisions regarding the use and deployment of AI models. But careful evaluations are needed to assess whether this expectation has been fulfilled. Current evaluations mainly focus on algorithmic properties of explanations, and those that involve human subjects often employ subjective questions to test human's perception of explanation usefulness, without being grounded in objective metrics and measurements. In this work, we evaluate whether explanations can improve human decision-making in practical scenarios of machine learning model development. We conduct a mixed-methods user study involving image data to evaluate saliency maps generated by SmoothGrad, GradCAM, and an oracle explanation on two tasks: model selection and counterfactual simulation. To our surprise, we did not find evidence of significant improvement on these tasks when users were provided with any of the saliency maps, even the synthetic oracle explanation designed to be simple to understand and highly indicative of the answer. Nonetheless, explanations did help users more accurately describe the models. These findings suggest caution regarding the usefulness and potential for misunderstanding in saliency-based explanations. | Evaluating the Utility of Model Explanations for Model Development | null | Workshop/ATTRIB | 2312.06032 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=wyalhhLXCa | @inproceedings{
anonymous2023the,
title={The Importance of Prompt Tuning for Automated Neuron Explanations},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=wyalhhLXCa}
} | Recent advances have greatly increased the capabilities of large language models (LLMs), but our understanding of the models and their safety has not progressed as fast. In this paper we aim to understand LLMs deeper by studying their individual neurons. We build upon previous work showing large language models such as GPT-4 can be useful in explaining what each neuron in a language model does. Specifically, we analyze the effect of the prompt used to generate explanations and show that reformatting the explanation prompt in a more natural way can significantly improve neuron explanation quality and greatly reduce computational cost. We demonstrate the effects of our new prompts in three different ways, incorporating both automated and human evaluations. | The Importance of Prompt Tuning for Automated Neuron Explanations | null | Workshop/ATTRIB | 2310.06200 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=uSvN2oozRK | @inproceedings{
anonymous2023does,
title={Does It Know?: Probing for Uncertainty in Language Model Latent Beliefs},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=uSvN2oozRK}
} | Understanding a language model's beliefs about its truthfulness is crucial for building more trustworthy, factually accurate large language models. The recent method of Contrast-Consistent Search (CCS) measures this "latent belief" via a linear probe on intermediate activations of a language model, trained in an unsupervised manner to classify inputs as true or false. As an extension of CCS, we propose Uncertainty-detecting CCS (UCCS), which encapsulates finer-grained notions of truth, such as uncertainty or ambiguity. Concretely, UCCS teaches a probe, using only unlabeled data, to classify a model's latent belief on input text as true, false, or uncertain. We find that UCCS is an effective unsupervised-trained selective classifier, using its uncertainty class to filter out low-confidence truth predictions, leading to improved accuracy across a diverse set of models and tasks. To properly evaluate UCCS predictions of truth and uncertainty, we introduce a toy dataset, named Temporally Measured Events (TYMES), which comprises true or falsified facts, paired with timestamps, extracted from recent news articles from the past several years. TYMES can be combined with any language model's training cutoff date to systematically produce a subset of data beyond (literally, occurring after) the knowledge limitations of the model. TYMES serves as a valuable proof-of-concept for how we can benchmark uncertainty or time-sensitive world knowledge in language models, a setting which includes but extends beyond our UCCS evaluations. | Does It Know?: Probing and Benchmarking Uncertainty in Language Model Latent Beliefs | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=tiLbFR4bJW | @inproceedings{
anonymous2023attribution,
title={Attribution Patching Outperforms Automated Circuit Discovery},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=tiLbFR4bJW}
} | Automated interpretability research has recently attracted attention as a potential research direction that could scale explanations of neural network behavior to large models. Existing automated circuit discovery work applies activation patching to identify subnetworks responsible for solving specific tasks (circuits). In this work, we show that a simple method based on attribution patching outperforms all existing methods while requiring just two forward passes and a backward pass. We apply a linear approximation to activation patching to estimate the importance of each edge in the computational subgraph. Using this approximation, we prune the least important edges of the network. We survey the performance and limitations of this method, finding that averaged over all tasks our method has greater AUC from circuit recovery than other methods. | Attribution Patching Outperforms Automated Circuit Discovery | null | Workshop/ATTRIB | 2310.10348 | [
"https://github.com/aaquib111/acdcpp"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=nb3aaC9Hr8 | @inproceedings{
anonymous2023in,
title={In Search of a Data Transformation that Accelerates Neural Field Training},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=nb3aaC9Hr8}
} | Neural field is a special type of neural network that represents a single datum. We study whether we can speed up the training of such neural networks, by fitting a transformed version of the target datum; one can recover the original signal by inverting back the signal represented by the trained neural field. We empirically find that very simple data transformations, such as color inversion or random pixel shuffling, can substantially speed up or slow down the training. In particular, to our surprise, we observe that an image with randomly shuffled pixels can be fit much faster, despite having a very large frequency. | In Search of a Data Transformation that Accelerates Neural Field Training | null | Workshop/ATTRIB | 2311.17094 | [
"https://github.com/effl-lab/dt4neural-field"
] | https://huggingface.co/papers/2311.17094 | 3 | 2 | 0 | 4 | [] | [] | [
"lyunm1206/Interactive_Loss_Landscapes"
] | [] | [] | [
"lyunm1206/Interactive_Loss_Landscapes"
] | 1 | poster |
null | https://openreview.net/forum?id=lG08cfNhlc | @inproceedings{
anonymous2023automatic,
title={Automatic Discovery of Visual Circuits},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=lG08cfNhlc}
} | To date, most discoveries of subnetworks that implement human-interpretable computations in deep vision models have involved close study of single units and large amounts of human labor. We explore scalable methods for extracting the subgraph of a vision model’s computational graph that underlies a particular capability. In this paper, we formulate capabilities as mappings of human-interpretable visual concepts to intermediate feature representations. We introduce a new method for identifying these subnetworks: specifying a visual concept using a few examples, and then tracing the interdependence of neuron activations across layers, or their functional connectivity. We find that our approach extracts circuits that causally affect model output, and that editing these circuits can defend large pretrained models from adversarial attacks. | Automatic Discovery of Visual Circuits | null | Workshop/ATTRIB | 2404.14349 | [
"https://github.com/multimodal-interpretability/visual-circuits"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=lDeysxpH6W | @inproceedings{
anonymous2023mining,
title={Mining the Diamond Miner: Mechanistic Interpretability on the Video PreTraining Agent},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=lDeysxpH6W}
} | Although decision-making systems based on reinforcement learning (RL) can be widely used in a variety of applications, their lack of interpretability raises concerns, especially in high-stakes scenarios. In contrast, Mechanistic Interpretability (MI) has shown potential in breaking down complex deep neural networks into understandable components in language and vision tasks. Accordingly, in this study, we apply MI to understand the behavior of a Video PreTraining (VPT) agent, exhibiting human-level proficiency in numerous Minecraft tasks. Our exploration is centered on the task of diamond mining and its associated subtasks, such as crafting wooden logs and iron pickaxes. By employing circuit analysis, we aim to decode the network's representation of these tasks and subtasks. We find a significant head in the VPT model encoding for an attacking action, although its ablation doesn't markedly affect the agent's performance. Our findings indicate that this approach can provide useful insights into the agent's behavior. | Mining the Diamond Miner: Mechanistic Interpretability on the Video PreTraining Agent | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=klVoedPgtY | @inproceedings{
anonymous2023threshold,
title={Threshold {KNN}-Shapley: A Linear-Time and Privacy-Friendly Approach to Data Valuation},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=klVoedPgtY}
} | Data valuation aims to quantify the usefulness of individual data sources in training machine learning (ML) models, and is a critical aspect of data-centric ML research. However, data valuation faces significant yet frequently overlooked privacy challenges despite its importance. This paper studies these challenges with a focus on KNN-Shapley, one of the most practical data valuation methods nowadays. We first emphasize the inherent privacy risks of KNN-Shapley, and demonstrate the significant technical difficulties in adapting KNN-Shapley to accommodate differential privacy (DP). To overcome these challenges, we introduce \emph{TKNN-Shapley}, a refined variant of KNN-Shapley that is privacy-friendly, allowing for straightforward modifications to incorporate DP guarantee (\emph{DP-}TKNN-Shapley). We show that DP-TKNN-Shapley has several advantages and offers a superior privacy-utility tradeoff compared to naively privatized KNN-Shapley in discerning data quality. Moreover, even non-private TKNN-Shapley achieves comparable performance as KNN-Shapley. Overall, our findings suggest that TKNN-Shapley is a promising alternative to KNN-Shapley, particularly for real-world applications involving sensitive data. | Threshold KNN-Shapley: A Linear-Time and Privacy-Friendly Approach to Data Valuation (Workshop Version) | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=k4cSPcBxX0 | @inproceedings{
anonymous2023how,
title={How Does Colour and Shape Goal Misgeneralization Happen in Reinforcement Learning},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=k4cSPcBxX0}
} | We explore colour versus shape goal misgeneralization originally demonstrated by Di Langosco et al. (2022) in the Procgen Maze environment, where, given an ambiguous choice, the agents seem to prefer generalization based on colour rather than shape. After training over 1,000 agents in a simplified version of the environment and evaluating them on over 10 million episodes, we conclude that the behaviour can be attributed to the agents learning to detect the goal object through a specific colour channel. This choice is arbitrary. Additionally, we show how, due to underspecification, the preferences can change when retraining the agents using exactly the same procedure except for using a different random seed for the training run. Finally, we demonstrate the existence of outliers in out-of-distribution behaviour based on training random seed alone. | Colour versus Shape Goal Misgeneralization in Reinforcement Learning: A Case Study | null | Workshop/ATTRIB | 2312.03762 | [
"https://github.com/KarolisRam/colour-shape-goal-misgeneralization"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=ie6hRGMCp2 | @inproceedings{
anonymous2023adversarial,
title={Adversarial Attacks on Neuron Interpretation via Activation Maximization},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=ie6hRGMCp2}
} | Feature visualization is one of the most popular techniques to interpret the internal behavior of individual units of trained deep neural networks. Based on activation maximization, they consist of finding $\textit{synthetic}$ or $\textit{natural}$ inputs that maximize neuron activations. This paper introduces an optimization framework that aims to deceive feature visualization through adversarial model manipulation. It consists of fine-tuning a pre-trained model with a specifically introduced loss that aims to maintain model performance, while also significantly changing feature visualization. We provide evidence of the success of this manipulation on several pre-trained models for the ImageNet classification task. | Adversarial Attacks on Neuron Interpretation via Activation Maximization | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=i0x5EKSDdq | @inproceedings{
anonymous2023divergence,
title={Divergence at the Interpolation Threshold: Identifying, Interpreting {\textbackslash}\& Ablating the Sources of a Deep Learning Puzzle},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=i0x5EKSDdq}
} | Machine learning models misbehave, often in unexpected ways. One prominent misbehavior is when the test loss diverges at the interpolation threshold, perhaps best known from its distinctive appearance in double descent. While considerable theoretical effort has gone into understanding generalization of overparameterized models, less effort has been made at understanding why the test loss misbehaves at the interpolation threshold. Moreover, analytically solvable models in this area employ a range of assumptions and use complex techniques from random matrix theory, statistical mechanics, and kernel methods, making it difficult to assess when and why test error might diverge. In this work, we analytically study the simplest supervised model - ordinary linear regression - and show intuitively and rigorously when and why a divergence occurs at the interpolation threshold using basic linear algebra. We identify three interpretable factors that, when all present, cause the divergence. We demonstrate on real data that linear models' test losses diverge at the interpolation threshold and that the divergence disappears when we ablate any one of the three identified factors. We conclude with insights on recent discoveries in nonlinear models regarding superposition and double descent. | Divergence at the Interpolation Threshold: Identifying, Interpreting Ablating the Sources of a Deep Learning Puzzle | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=hGtjFfO27f | @inproceedings{
anonymous2023the,
title={The Reversal Curse: {LLM}s trained on ''A is B'' fail to learn ''B is A''},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=hGtjFfO27f}
} | We expose a surprising failure of generalization in auto-regressive large language models (LLMs). If a model is trained on a sentence of the form "*A is B*", it will not automatically generalize to the reverse direction "*B is A*". This is the **Reversal Curse**. For instance, if a model is trained on "Olaf Scholz was the ninth Chancellor of Germany", it will not automatically be able to answer the question, "Who was the ninth Chancellor of Germany?". Moreover, the likelihood of the correct answer ("Olaf Scholz") will not be higher than for a random name. Thus, models exhibit a basic failure of logical deduction and do not generalize a prevalent pattern in their training set (i.e. if "*A is B*" occurs, "*B is A*" is more likely to occur). We provide evidence for the Reversal Curse by finetuning GPT-3 and Llama-1 on fictitious statements such as "Uriah Hawthorne is the composer of *Abyssal Melodies*" and showing that they fail to correctly answer "Who composed *Abyssal Melodies?*". The Reversal Curse is robust across model sizes and model families and is not alleviated by data augmentation. We also evaluate ChatGPT (GPT-3.5 and GPT-4) on questions about real-world celebrities, such as "Who is Tom Cruise's mother? [A: Mary Lee Pfeiffer]" and the reverse "Who is Mary Lee Pfeiffer's son?". GPT-4 correctly answers questions like the former 79% of the time, compared to 33% for the latter. This shows a failure of logical deduction that we hypothesize is caused by the Reversal Curse. | The Reversal Curse: LLMs trained on "A is B" fail to learn "B is A" | null | Workshop/ATTRIB | 2309.12288 | [
"https://github.com/lukasberglund/reversal_curse"
] | https://huggingface.co/papers/2309.12288 | 2 | 3 | 0 | 7 | [] | [
"lberglund/reversal_curse"
] | [] | [] | [
"lberglund/reversal_curse"
] | [] | 1 | poster |
null | https://openreview.net/forum?id=giMJzZIuzr | @inproceedings{
anonymous2023the,
title={The Geometry of Truth: Emergent Linear Structure in Large Language Model Representations of True/False Datasets},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=giMJzZIuzr}
} | Large Language Models (LLMs) have impressive capabilities, but are also prone to outputting falsehoods. Recent work has developed techniques for inferring whether a LLM is telling the truth by training probes on the LLM's internal activations. However, this line of work is controversial, with some authors pointing out failures of these probes to generalize in basic ways, among other conceptual issues. In this work, we curate high-quality datasets of true/false statements and use them to study in detail the structure of LLM representations of truth, drawing on three lines of evidence: 1. Visualizations of LLM true/false statement representations, which reveal clear linear structure. 2. Transfer experiments in which probes trained on one dataset generalize to different datasets. 3. Causal evidence obtained by surgicallly intervening in a LLM's forward pass, causing it to treat false statements as true and vice versa. Overall, we present evidence that language models linearly represent the truth or falsehood of factual statements. We also introduce a novel technique, mass-mean probing, which generalizes better and is more causally implicated in model outputs than other probing techniques. | The Geometry of Truth: Emergent Linear Structure in Large Language Model Representations of True/False Datasets | null | Workshop/ATTRIB | 2310.06824 | [
"https://github.com/saprmarks/geometry-of-truth"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=csilNwWnoG | @inproceedings{
anonymous2023efficient,
title={Efficient Data Valuation for Weighted Nearest Neighbor Algorithms},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=csilNwWnoG}
} | Data Shapley is a principled way to assess the importance of individual training data sources for machine learning (ML) applications. However, it often comes with computational challenges in calculating exact Data Shapley scores. KNN-Shapley \citep{jia2019efficient}, which assigns data value leveraging the efficiently computable Data Shapley score of $K$ nearest neighbors (KNN), has gained popularity as a viable alternative due to its computationally efficient nature. However, \cite{jia2019efficient} only gives a practical algorithm for computing Data Shapley for unweighted KNN, but weighted KNN is more prevalently used in practice.
This work addresses the computational challenges of calculating the exact Data Shapley for weighted KNN classifiers (WKNN-Shapley). By making small adjustments to KNN configurations, we recast the computation of WKNN-Shapley into a counting problem and introduce an $O(K^2 N^2)$ algorithm, presenting a notable improvement from the naive, impractical $O(N^K)$ algorithm. We also develop a deterministic approximation algorithm that further improves computational efficiency while maintaining the key fairness properties of the Shapley value. These advancements position WKNN-Shapley as a compelling alternative to KNN-Shapley. In particular, WKNN-Shapley can select high-quality data points and improve the performance of retrieval-augmented language models. | Efficient Data Valuation for Weighted Nearest Neighbor Algorithms | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=cGui9S5GqL | @inproceedings{
anonymous2023is,
title={Is This the Subspace You Are Looking for? An Interpretability Illusion for Subspace Activation Patching},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=cGui9S5GqL}
} | Mechanistic interpretability aims to understand model behaviors in terms of specific, interpretable features, often hypothesized to manifest as low-dimensional subspaces of activations.
Specifically, recent studies have explored subspace interventions (such as activation patching) as a way to both manipulate model behavior and attribute the features behind it to given subspaces.
In this work, we demonstrate that these two aims diverge, potentially leading to an illusory sense of interpretability.
Counterintuitively, even if a subspace intervention modifies end-to-end model behavior in the desired way, this effect may be achieved by activating a \emph{dormant parallel pathway} leveraging a component that is \emph{causally disconnected} from model outputs.
We demonstrate this phenomenon in a distilled mathematical example, in two real-world domains (the indirect object identification task and factual recall), and present evidence for its prevalence in practice.
In the context of factual recall, we further show a link to rank-1 fact editing, providing a mechanistic explanation for previous work observing an inconsistency between fact editing performance and fact localization.
Finally, we remark on what a success case of subspace activation patching looks like. | Is This the Subspace You Are Looking for? An Interpretability Illusion for Subspace Activation Patching | null | Workshop/ATTRIB | 2311.17030 | [
"https://github.com/amakelov/activation-patching-illusion"
] | https://huggingface.co/papers/2311.17030 | 3 | 0 | 0 | 3 | [] | [] | [] | [] | [] | [] | 1 | poster |
null | https://openreview.net/forum?id=bqOzmRPGCX | @inproceedings{
anonymous2023object,
title={Object Detection in Deep Neural Networks Differs from Humans in the Periphery},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=bqOzmRPGCX}
} | To understand how strategies used by object detection models compare to those in human vision, we simulate peripheral vision in object detection models at the input stage. We collect human data on object change detection in the periphery and compare it to detection models with a simulated periphery. We find that unlike humans, models are highly sensitive to the texture-like transformation in peripheral vision. Not only do models under-perform compared to humans, they do not follow the same clutter effects as humans even when fixing the model task to closely mimic the human one. Training on peripheral input boosts performance on the change detection task, but appears to aid object localization in the periphery much more than object identification. This suggests that human-like performance is not attributable to input data alone, and to fully address the differences we see in human and model detection, farther downstream changes may be necessary. In the future, improving alignment between object detection models and human representations could help us build models with more human-explainable detection strategies. | Object Detection in Deep Neural Networks Differs from Humans in the Periphery | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=Zqj3YQ1QAC | @inproceedings{
anonymous2023formal,
title={Formal Definition of Fingerprints Improves Attribution of Generative Models},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=Zqj3YQ1QAC}
} | Recent works have shown that generative models leave traces of their underlying
generative process on the generated samples, broadly referred to as fingerprints of a
generative model, and have studied their utility in detecting synthetic images from
real ones. However, the extent to which these fingerprints can distinguish between
various types of synthetic images and help identify the underlying generative
process remain under-explored. In particular, the very definition of a fingerprint
remains unclear, to our knowledge. To that end, in this work, we formalize the
definition of artifact and fingerprint in generative models, propose an algorithm for
computing them in practice, and finally study how different design parameters affect
the model fingerprints and their attributability. We find that using our proposed
definition can significantly improve the performance on the task of identifying
the underlying generative process from samples (model attribution) compared to
existing methods. Additionally, we study the structure of the fingerprints and
observe that it is very predictive of the effect of different design choices on the
generative process. | Formal Definition of Fingerprints Improves Attribution of Generative Models | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=ZP7hgFNXQy | @inproceedings{
anonymous2023attributing,
title={Attributing Learned Concepts in Neural Networks to Training Data},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=ZP7hgFNXQy}
} | By now there is substantial evidence that deep learning models learn certain human-interpretable features as part of their internal representations of data. As having the right (or wrong) concepts is critical to trustworthy machine learning systems, it is natural to ask which inputs from the model's original training set were most important for learning a concept at a given layer. To answer this, we combine data attribution methods with methods for probing the concepts learned by a model. Training network and probe ensembles for two concept datasets on a range of network layers, we use the recently developed TRAK method for large-scale data attribution. We find some evidence for *convergence*, where removing the 10,000 top attributing images for a concept and retraining the model does not change the location of the concept in the network nor the probing sparsity of the concept. This suggests that rather than being highly dependent on a few specific examples, the features that inform the development of a concept are spread in a more diffuse manner across its exemplars, implying robustness in concept formation. | Attributing Learned Concepts in Neural Networks to Training Data | null | Workshop/ATTRIB | 2310.03149 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=XUIYn3jo5T | @inproceedings{
anonymous2023when,
title={When Less is More: Investigating Data Pruning for Pretraining {LLM}s at Scale},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=XUIYn3jo5T}
} | Large volumes of text data have contributed significantly to the development of large language models (LLMs) in recent years. To date, efforts to prune these datasets to higher quality subsets have relied on hand-crafted heuristics encoded as rule-based filters. In this work, we explore scalable estimates of data quality that can be used to systematically measure the quality of pretraining data, namely perplexity, the Error L2-Norm, and memorization. These metrics are used to rank and prune pretraining corpora, and we subsequently compare LLMs trained on these pruned datasets. We find that perplexity outperforms other scoring methods and improves over our no-pruning baseline while training on as little as 30\% of the original training dataset. Our work sets a foundation for strategies in automatically curating high quality corpora and suggests that large amounts of pretraining data can be removed while retaining performance. | When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale | null | Workshop/ATTRIB | 2309.04564 | [
""
] | https://huggingface.co/papers/2309.04564 | 5 | 15 | 0 | 6 | [] | [] | [] | [] | [] | [] | 1 | poster |
null | https://openreview.net/forum?id=WGjQt3aDn7 | @inproceedings{
anonymous2023a,
title={A Simple and Efficient Baseline for Data Attribution on Images},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=WGjQt3aDn7}
} | Data attribution methods play a crucial role in understanding machine learning models, providing insight into which training data points are most responsible for model outputs during deployment. However, current state-of-the-art approaches require a large ensemble of as many as 300,000 models to accurately attribute model predictions. These approaches therefore come at a high computational cost, are memory intensive, and are hard to scale to large models or datasets.
In this work, we focus on a minimalist baseline that relies on the image features from a pretrained self-supervised backbone to retrieve images from the dataset. Our method is model-agnostic and scales easily to large datasets. We show results on CIFAR-10 and ImageNet, achieving strong performance that rivals or outperforms state-of-the-art approaches at a fraction of the compute or memory cost. Contrary to prior work, our results reinforce the intuition that a model's prediction on one image is most impacted by visually similar training samples. Our approach serves as a simple and efficient baseline for data attribution on images. | A Simple and Efficient Baseline for Data Attribution on Images | null | Workshop/ATTRIB | 2311.03386 | [
"https://github.com/vasusingla/simple-data-attribution"
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=LMV6HuZJ2L | @inproceedings{
anonymous2023exploring,
title={Exploring Dataset-Scale Indicators of Data Quality},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=LMV6HuZJ2L}
} | Modern computer vision foundation models are trained on massive amounts of data, incurring large economic and environmental costs. Recent research has suggested that improving data quality can significantly reduce the need for data quantity. But what constitutes data quality in computer vision? We posit that the quality of a given dataset can be decomposed into distinct sample-level and dataset-level constituents, and that the former have been more extensively studied than the latter. We ablate the effects of two important dataset-level constituents: label set design, and class balance. By monitoring these constituents using key indicators we provide, researchers and practitioners can better anticipate model performance, measured in terms of its accuracy and robustness to distribution shifts. | Exploring Dataset-Scale Indicators of Data Quality | null | Workshop/ATTRIB | 2311.04016 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=KLVL0Kj6E8 | @inproceedings{
anonymous2023selfselect,
title={Self-Select: Optimizing Instruction Selection for Large Language Models},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=KLVL0Kj6E8}
} | The same question can often be presented in different ways, depending on the audience and the intent with which it is being posed. To determine whether large language models (LLMs) demonstrate preferences for one phrasing over another regardless of semantic content, we introduce \textit{Self-Select}, a method for selection of a preferred instruction template, and generation of high-quality synthetic data samples. This algorithm makes use of a \textit{meta-prompt} to decide on an instruction template, given a task and candidate templates then generates $n$ new samples using the chosen template. We evaluate \textit{Self-Select} on numerical reasoning and sentiment classification tasks, using a variety of instruction-tuned and base models, providing insights into their abilities and biases. We find that permuting the instruction template ordering in the prompt leads to vastly different choice distributions, suggesting that selections of a specific template can be attributed to inductive biases rather than semantic understanding, even after instruction-tuning. | Self-Select: Optimizing Instruction Selection for Large Language Models | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=Hi9zCFiiNw | @inproceedings{
anonymous2023speculative,
title={Speculative Behavior: An Approach to Large Language Model Evaluation and Optimization},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=Hi9zCFiiNw}
} | Trained Large Language Models (LLMs) have gained significant interest due to their ability to interpret natural language instructions and address a wide range of tasks with high proficiency. However, in practice, these models pose multiple challenges. On one hand, it is exceedingly difficult to control and ensure that the model's behavior remains consistent, harmless, and safe. On the other hand, the most advanced models are delivered via APIs as black-box services, making it challenging to guarantee their proper behavior. Addressing these challenges has become an urgent concern, especially in environments where a model's response can impact safety and trustworthiness. Many recent studies focus on the evaluation of models using benchmarks based on community-curated datasets. However, this form of evaluation is prone to data leakage and premature dataset obsolescence. Moreover, it doesn't necessarily align with all the specific goals that may be desired. One alternative for aligning specific objectives with the model behavior is fine-tuning, but this process is time-consuming and might be prohibitively expensive for many organizations. In this study, we propose the idea of measuring the model's behavior towards specific objectives through the concept of Speculative Behavior Equivalence (SBE). We introduce a general, agnostic approach that can be adapted to various models and tailored to the unique metrics of individual cases whilst remaining constrained to specific budgets. Additionally, we formulate the Speculative Behavior-Based Optimization problem (CSBO), which presents an opportunity to leverage AutoML techniques in the field of LLMs for optimizing behavior. | Speculative Behavior: An Approach to Large Language Model Evaluation and Optimization | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=HVyDhCc7LW | @inproceedings{
anonymous2023large,
title={Large Language Model Attributions: A Unified Perspective},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=HVyDhCc7LW}
} | As businesses, products, and services spring up around large language models, the trustworthiness of these models hinges on the verifiability of their outputs. However, methods for explaining language model outputs largely fall across two distinct fields of study which both use the term "attribution" to refer to entirely separate techniques: citation generation and training data attribution. In many modern applications, such as legal document generation and medical question answering, both types of attributions are important. In this work, we argue for and present a unified framework of large language model attributions. We show how existing methods of different types of attribution fall under the unified framework. We also use the framework to discuss real-world use cases where one or both types of attributions are required. We believe that this unified framework will guide the use case driven development of systems that leverage both types of attribution, as well as the standardization of their evaluation. | Unifying Corroborative and Contributive Attributions in Large Language Models | null | Workshop/ATTRIB | 2311.12233 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | oral |
|
null | https://openreview.net/forum?id=G8RRE6zDGm | @inproceedings{
anonymous2023algorithm,
title={Algorithm Selection with Priority Order for Instances},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=G8RRE6zDGm}
} | Reliability in medical image diagnostics is a required trait for any artificial system. Currently, most approaches rely on highly trained and specific models to leverage the feature quality learned from a particular type of medium, such as X-rays, NMR, PET scans and others. While this approach aligns with the standard human expert perspective, it also limits artificial systems to the representations learned from the dataset distribution. To gain a better understanding of how different media affect specific tasks, we explore task-specific feature transfer between domains. In this work, we propose the possibility of merging features from various areas to harness feature transfer in outlier cases. For this purpose, we develop an Algorithm Selection (AS) method that chooses algorithms trained on different sets of medical images and for different classification tasks. The AS system is then applied to a different classification task. The AS represents a set of methods that, given a problem and a range of existing algorithms, selects the best algorithm on a case- by-case basis. The results demonstrate the advantages of incorporating algorithms from different tasks and datasets in a supervised manner. By considering algorithms trained on diverse datasets, we can effectively capture outliers that might otherwise be neglected by more specific algorithms. | Algorithm Selection with Priority Order for Instances | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
||
null | https://openreview.net/forum?id=Ei4k9xurmC | @inproceedings{
anonymous2023prototype,
title={Prototype Generation: Robust Feature Visualisation for Data Independent Interpretability},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=Ei4k9xurmC}
} | We introduce Prototype Generation, a stricter and more robust form of feature visualisation for model-agnostic, data-independent interpretability of image classification models. We demonstrate its ability to generate inputs that result in natural activation paths, countering previous claims that feature visualisation algorithms are untrustworthy due to the unnatural internal activations. We substantiate these claims by quantitatively measuring similarity between the internal activations of our generated prototypes and natural images. We also demonstrate how the interpretation of generated prototypes yields important insights, highlighting spurious correlations and biases learned by models which quantitative methods over test-sets cannot identify. | Prototype Generation: Robust Feature Visualisation for Data Independent Interpretability | null | Workshop/ATTRIB | 2309.17144 | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
|
null | https://openreview.net/forum?id=EKvqw9k3lC | @inproceedings{
anonymous2023backtracking,
title={Backtracking Mathematical Reasoning of Language Models to the Pretraining Data},
author={Anonymous},
booktitle={NeurIPS Workshop on Attributing Model Behavior at Scale},
year={2023},
url={https://openreview.net/forum?id=EKvqw9k3lC}
} | In-context learning and chain-of-thought prompting have demonstrated surprising performance improvements on mathematical reasoning benchmarks.
Therefore, understanding the underlying factors enabling these capabilities is crucial.
However, the specific aspects of pretraining data that equip models with mathematical reasoning capabilities remain largely unexplored and are less studied systematically.
In this study, we identify subsets of model pretraining data that contribute to math reasoning ability of the model, and evaluate it on several mathematical operations (e.g. addition, multiplication) and tasks (e.g. the asdiv dataset).
We measure the importance of such subsets by continual training of the model on pretraining data subsets, and then we quantify the change in performance on the mathematical benchmark to assess their importance.
If a subset results in an improved performance, we conjecture that such subset contributes to a model's overall mathematical ability.
Our results unveil that while training on math-only data contributes to simple arithmetic abilities, it does not solely explain performance on more complex reasoning abilities like chain-of-thought reasoning.
We also find that code data contributes to chain-of-thought reasoning while reducing the arithmetic performance. | Backtracking Mathematical Reasoning of Language Models to the Pretraining Data | null | Workshop/ATTRIB | [
""
] | -1 | -1 | -1 | -1 | [] | [] | [] | [] | [] | [] | 0 | poster |
Subsets and Splits