Haggai Maron

Haggai Maron

Assistant Professor · Technion, Faculty of Electrical and Computer Engineering
Senior Research Scientist · NVIDIA Research, Tel Aviv

My primary research interest is in machine learning, with a focus on deep learning for structured data. Specifically, I study how to apply deep learning techniques to sets, graphs, point clouds, surfaces, weight spaces and other mathematical objects that have an inherent symmetry structure. My goal is twofold: first, to understand and design deep learning architectures from a theoretical perspective, for example, by analyzing their expressive power; and second, to demonstrate their practical effectiveness on real-world problems involving structured data.

I completed my Ph.D. in 2019 at the Weizmann Institute of Science under the supervision of Prof. Yaron Lipman.

Email: haggaimaron (at) technion.ac.il  ·  Google Scholar

Group

Current Members Fabrizio Frasca (Postdoc, 2024–today) · Yam Eitan (PhD, 2024–today) · Guy Bar-Shalom (PhD, 2023–today, joint w/ Ran El-Yaniv) · Yoav Gelberg (PhD, 2024–today, joint w/ Michael Bronstein) · Adir Dayan (PhD, 2025–today) · Alina Sudakov (MSc, 2026–today) · Yonathan Wolanowsky (MSc, 2026–today) Past Members Ran Elbaz (MSc, 2023–2025) · Yaniv Galron (MSc, 2023–2025, joint w/ Eran Treister) · Yuval Aidan (MSc, 2023–2025, joint w/ Ayellet Tal) · Edan Kinderman (MSc, 2023–2025, joint w/ Daniel Soudry) · Ofir Haim (MSc, 2023–2025, joint w/ Shie Mannor) Close Collaborators Theo (Moe) Putterman (UC Berkeley, 2023–2025) · Beatrice Bevilacqua (Purdue, 2021–2025) · Derek Lim (MIT CSAIL, 2021–2025) · Moshe Eliasof (Cambridge/BGU, 2022–today) · Aviv Navon (BIU/Aiola, 2021–today) · Aviv Shamsian (BIU/Aiola, 2022–today)

News

Teaching

  • 2025/Spring (Technion): Introduction to Machine Learning
  • 2025/Spring (Technion): Deep Learning and Groups
  • 2024/Winter (Technion): Topics in Learning on Graphs
  • 2024/Spring (Technion): Deep Learning and Groups
  • 2023/Winter (Technion): Topics in Learning on Graphs
  • 2019/Spring (WIS): Geometric and Algebraic Methods in Deep Learning
  • 2018/Winter (WIS): Geometry and Deep Learning

Publications

CHARM

Neural Message-Passing on Attention Graphs for Hallucination Detection

The Fourteenth International Conference on Learning Representations (ICLR 2026)

Summary

Large Language Models (LLMs) often generate incorrect or unsupported content, known as hallucinations. Existing detection methods rely on heuristics or simple models over isolated computational traces such as activations, or attention maps. We unify these signals by representing them as attributed graphs, where tokens are nodes, edges follow attentional flows, and both carry features from attention scores and activations. Our approach, CHARM, casts hallucination detection as a graph learning task and tackles it by applying GNNs over the above attributed graphs. We show that CHARM provably subsumes prior attention-based heuristics and, experimentally, it consistently outperforms other leading approaches across diverse benchmarks.

Paper OpenReview

FS-KAN

FS-KAN: Permutation Equivariant Kolmogorov-Arnold Networks via Function Sharing

Ran Elbaz, Guy Bar-Shalom, Yam Eitan, Fabrizio Frasca, Haggai Maron

The Fourteenth International Conference on Learning Representations (ICLR 2026)

Summary

This paper introduces Function Sharing KAN (FS-KAN), a principled approach to constructing equivariant and invariant KA layers for arbitrary permutation symmetry groups, unifying and significantly extending previous work in this domain. We derive the basic construction of these FS-KAN layers by generalizing parameter-sharing schemes to the Kolmogorov-Arnold setup and provide a theoretical analysis demonstrating that FS-KANs have the same expressive power as networks that use standard parameter-sharing layers. Empirical evaluations show that FS-KANs exhibit superior data efficiency compared to standard parameter-sharing layers, making them an excellent architecture choice in low-data regimes.

Paper OpenReview

HOD-GNN

On The Expressive Power of GNN Derivatives

Yam Eitan, Moshe Eliasof, Yoav Gelberg, Fabrizio Frasca, Guy Bar-Shalom, Haggai Maron

The Fourteenth International Conference on Learning Representations (ICLR 2026)

Summary

We introduce High-Order Derivative GNN (HOD-GNN), a novel method that enhances the expressivity of Message Passing Neural Networks (MPNNs) by leveraging high-order node derivatives of the base model. These derivatives generate expressive structure-aware node embeddings processed by a second GNN in an end-to-end trainable architecture. Theoretically, we show that the resulting architecture family's expressive power aligns with the WL hierarchy. We also draw deep connections between HOD-GNN, Subgraph GNNs, and popular structural encoding schemes.

Paper

WS-KAN

A Graph Meta-Network for Learning on Kolmogorov–Arnold Networks

Guy Bar-Shalom, Ami Tavory, Itay Evron, Maya Bechler-Speicher, Ido Guy, Haggai Maron

The Fourteenth International Conference on Learning Representations (ICLR 2026)

Summary

We show that KANs share the same permutation symmetries as MLPs, and propose the KAN-graph, a graph representation of their computation. Building on this, we develop WS-KAN, the first weight-space architecture that learns on KANs, which naturally accounts for their symmetry. We analyze WS-KAN's expressive power and construct a comprehensive 'zoo' of trained KANs as benchmarks. Across all tasks, WS-KAN consistently outperforms structure-agnostic baselines, often by a substantial margin.

OpenReview

HistoGraph

Learning from Historical Activations in Graph Neural Networks

Yaniv Galron, Hadar Sinai, Haggai Maron, Moshe Eliasof

The Fourteenth International Conference on Learning Representations (ICLR 2026)

Summary

We introduce HistoGraph, a novel two-stage attention-based final aggregation layer that first applies a unified layer-wise attention over intermediate activations, followed by node-wise attention. By modeling the evolution of node representations across layers, our HistoGraph leverages both the activation history of nodes and the graph structure to refine features used for final prediction. Empirical results demonstrate that HistoGraph offers strong performance that consistently improves traditional techniques, with particularly strong robustness in deep GNNs.

Paper

LOS

Beyond Next Token Probabilities: Learnable, Fast Detection of Hallucinations and Data Contamination on LLM Output Distributions

Guy Bar-Shalom*, Fabrizio Frasca*, Derek Lim, Yoav Gelberg, Yftah Ziser, Ran El-Yaniv, Gal Chechik, Haggai Maron

The 40th Annual AAAI Conference on Artificial Intelligence (AAAI 2026)

Spotlight at ICLR 2025 workshop 'Quantify Uncertainty and Hallucination in Foundation Models'

Summary

We propose that gray-box analysis should leverage the complete observable output of LLMs, consisting of both token probabilities and the complete token distribution sequences – a unified data type we term LOS (LLM Output Signature). We develop a transformer-based approach to process LOS that theoretically guarantees approximation of existing techniques while enabling more nuanced analysis. Our approach achieves superior performance on hallucination and data contamination detection in gray-box settings, significantly outperforming existing baselines.

Paper Code

LoL

GL Equivariant Metanetworks for Learning on Low Rank Weight Spaces

Theo Putterman, Derek Lim, Yoav Gelberg, Michael M. Bronstein, Stefanie Jegelka, Haggai Maron

The Fourth Learning on Graphs Conference (LOG 2025)

Oral Presentation

Summary

We investigate the potential of Learning on LoRAs (LoL), a setup where machine learning models learn and make predictions on datasets of LoRA weights. We first identify the inherent parameter symmetries of low-rank decompositions of weights and develop several symmetry-aware invariant or equivariant LoL models. In diverse experiments, we show that our LoL architectures can process LoRA weights to predict CLIP scores, finetuning data attributes, and accuracy on downstream tasks over 50,000 times faster than standard evaluation.

Paper

ACT-ViT

Beyond Token Probes: Hallucination Detection via Activation Tensors with ACT-ViT

Guy Bar-Shalom, Fabrizio Frasca, Yaniv Galron, Yftah Ziser, Haggai Maron

39th Annual Conference on Neural Information Processing Systems (NeurIPS 2025)

Summary

We introduce ACT-ViT, a Vision Transformer-inspired model that can be effectively and efficiently applied to activation tensors and supports training on data from multiple LLMs simultaneously. Through comprehensive experiments encompassing diverse LLMs and datasets, we demonstrate that ACT-ViT consistently outperforms traditional probing techniques while remaining extremely efficient for deployment. Our architecture benefits substantially from multi-LLM training, achieves strong zero-shot performance on unseen datasets, and can be transferred effectively to new LLMs through fine-tuning.

Paper

GradMetaNet

GradMetaNet: An Equivariant Architecture for Learning on Gradients

Yoav Gelberg, Yam Eitan, Aviv Navon, Aviv Shamsian, Theo Putterman, Michael M. Bronstein, Haggai Maron

39th Annual Conference on Neural Information Processing Systems (NeurIPS 2025)

Summary

We present a principled approach for designing architectures that process gradients guided by three principles: (1) equivariant design that preserves neuron permutation symmetries, (2) processing sets of gradients across multiple data points to capture curvature information, and (3) efficient gradient representation through rank-1 decomposition. We introduce GradMetaNet, a novel architecture for learning on gradients, and prove universality results. We demonstrate GradMetaNet's effectiveness on diverse gradient-based tasks for MLPs and transformers, such as learned optimization, INR editing, and loss landscape curvature estimation.

Paper

Temporal Laplacian PE

Understanding and Improving Laplacian Positional Encodings For Temporal GNNs

Yaniv Galron, Fabrizio Frasca, Haggai Maron, Eran Treister, Moshe Eliasof

ECML-PKDD 2025

Summary

We address key challenges in supra-Laplacian positional encodings for temporal graphs: high eigendecomposition costs, limited theoretical understanding, and ambiguity about when to apply these encodings. We offer a theoretical framework connecting supra-Laplacian encodings to per-time-slice encodings, introduce novel methods to reduce computational overhead (achieving up to 56x faster runtimes), and conduct an extensive experimental study to identify which models, tasks, and datasets benefit most from these encodings.

Paper

HyMN

Balancing Efficiency and Expressiveness: Subgraph GNNs with Walk-Based Centrality

Joshua Southern*, Yam Eitan, Guy Bar-Shalom, Michael Bronstein, Haggai Maron, Fabrizio Frasca*

International Conference on Machine Learning (ICML) 2025

Summary

We propose HyMN, an expressive and efficient approach that combines Subgraph GNNs and Structural Encodings (SEs) via walk-based centrality measures. By drawing a connection to perturbation analysis, we highlight the effectiveness of centrality-based sampling, which significantly reduces the computational burden of Subgraph GNNs. HyMN effectively addresses the expressiveness limitations of MPNNs while mitigating the computational costs of Subgraph GNNs, outperforming other subgraph sampling approaches while being competitive with full-bag Subgraph GNNs.

Paper Code

TDL

Topological Blind Spots: Understanding and Extending Topological Deep Learning Through the Lens of Expressivity

Yam Eitan, Yoav Gelberg, Guy Bar-Shalom, Fabrizio Frasca, Michael Bronstein, Haggai Maron

International Conference on Learning Representations (ICLR) 2025

Oral Presentation

Summary

We aim to explore both the strengths and weaknesses of higher-order message-passing (HOMP)'s expressive power and subsequently design novel architectures to address its limitations. We demonstrate HOMP's inability to distinguish between topological objects based on fundamental topological and metric properties such as diameter, orientability, planarity, and homology. We then develop two new classes of TDL models: multi-cellular networks (MCN) and scalable multi-cellular networks (SMCN), which mitigate many of HOMP's expressivity limitations.

Paper

Spectral Homomorphism

Homomorphism Expressivity of Spectral Invariant Graph Neural Networks

Jingchu Gai, Yiheng Du, Bohang Zhang, Haggai Maron, Liwei Wang

International Conference on Learning Representations (ICLR) 2025

Oral Presentation

Summary

We address the fundamental question of the power of spectral invariants in GNNs through the lens of homomorphism expressivity, providing a comprehensive and quantitative analysis. We prove that spectral invariant GNNs can homomorphism-count exactly a class of specific tree-like graphs referred to as parallel trees, and highlight the significance of this result in establishing quantitative expressiveness hierarchies across different architectural variants.

Paper

Lightning-Fast Inversion

Lightning-Fast Image Inversion and Editing for Text-to-Image Diffusion Models

Dvir Samuel, Barak Meiri, Haggai Maron, Yoad Tewel, Nir Darshan, Shai Avidan, Gal Chechik, Rami Ben-Ari

International Conference on Learning Representations (ICLR) 2025

Summary

We formulate diffusion inversion as finding the roots of an implicit equation and design a method based on Newton-Raphson (NR). We derive an efficient guided formulation that fastly converges and provides high-quality reconstructions and editing. Our solution, Guided Newton-Raphson Inversion, inverts an image within 0.4 sec (on an A100 GPU) for few-step models (SDXL-Turbo and Flux.1), opening the door for interactive image editing.

Paper

Heat Kernels

Directed Graph Generation with Heat Kernels

Marc T Law, Karsten Kreis, Haggai Maron

TMLR 2025

Summary

We propose a denoising autoencoder-based generative model that exploits the global structure of directed graphs via their Laplacian dynamics and enables one-shot generation. Our approach generalizes a special class of exponential kernels over discrete structures, called diffusion kernels or heat kernels, to the non-symmetric case via Reproducing Kernel Banach Spaces. We demonstrate that our model is able to generate directed graphs that follow the distribution of the training dataset even if it is multimodal.

Paper

Foldable Supernets

Foldable Supernets: Scalable Merging of Transformers with Different Initializations and Tasks

Edan Kinderman, Itay Hubara, Haggai Maron, Daniel Soudry

TMLR 2025

Summary

We propose Foldable SuperNet Merge (FS-Merge), a method that optimizes a SuperNet to fuse original models using a feature reconstruction loss. FS-Merge is simple, data-efficient, and capable of merging models of varying widths. We test FS-Merge against existing methods on MLPs and transformers across various settings, sizes, tasks, and modalities. FS-Merge consistently outperforms them, achieving SOTA results, particularly in limited data scenarios.

Paper

Graph Coarsening Subgraph GNN

A Flexible, Equivariant Framework for Subgraph GNNs via Graph Products and Graph Coarsening

Guy Bar-Shalom, Yam Eitan, Fabrizio Frasca, Haggai Maron

38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024)

Best Paper Award at Symmetry and Geometry in Neural Representations Workshop @ NeurIPS 2024

Summary

We introduce a new Subgraph GNNs framework employing a graph coarsening function to cluster nodes into super-nodes. The product between the coarsened and original graph reveals an implicit structure whereby subgraphs are associated with specific sets of nodes. By running generalized message-passing on such graph product, our method effectively implements an efficient yet powerful Subgraph GNN. We discover that the resulting node feature tensor exhibits new, unexplored permutation symmetries and leverage this structure to characterize the associated linear equivariant layers.

Paper

TracksTo4D

Fast Encoder-Based 3D from Casual Videos via Point Track Processing

Yoni Kasten, Wuyue Lu, Haggai Maron

38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024)

Summary

We present TracksTo4D, a learning-based approach that enables inferring 3D structure and camera positions from dynamic content originating from casual videos using a single efficient feed-forward pass. We propose operating directly over 2D point tracks as input and designing an architecture tailored for processing 2D point tracks. TracksTo4D is trained in an unsupervised way and can reconstruct a temporal point cloud and camera positions with accuracy comparable to state-of-the-art methods, while drastically reducing runtime by up to 95%.

Paper

GRANOLA

GRANOLA: Adaptive Normalization for Graph Neural Networks

Moshe Eliasof, Beatrice Bevilacqua, Carola-Bibiane Schönlieb, Haggai Maron

38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024)

Summary

We propose GRANOLA, a novel graph-adaptive normalization layer. Unlike existing normalization layers, GRANOLA normalizes node features by adapting to the specific characteristics of the graph, particularly by generating expressive representations of its neighborhood structure via Random Node Features (RNF). Our extensive empirical evaluation underscores the superior performance of GRANOLA over existing normalization techniques, emerging as the top-performing method within the same time complexity of MPNNs.

Paper

Parameter Symmetries

The Empirical Impact of Neural Parameter Symmetries, or Lack Thereof

Derek Lim*, Theo (Moe) Putterman*, Robin Walters, Haggai Maron, Stefanie Jegelka

38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024)

Best Paper Award at HiLD 2024: 2nd Workshop on High-dimensional Learning Dynamics @ ICML 2024

Summary

We empirically investigate the impact of neural parameter symmetries by introducing new neural network architectures that have reduced parameter space symmetries. We develop two methods of modifying standard neural networks to reduce parameter space symmetries and conduct a comprehensive experimental study assessing the effect of removing these symmetries. Our experiments reveal several interesting observations; for instance, we observe linear mode connectivity and monotonic linear interpolation in our networks, without any alignment of weight spaces.

Paper Workshop Paper

Graph Foundation Models

Towards Foundation Models on Graphs: An Analysis on Cross-Dataset Transfer of Pretrained GNNs

Fabrizio Frasca, Fabian Jogl, Moshe Eliasof, Matan Ostrovsky, Carola-Bibiane Schönlieb, Thomas Gärtner, Haggai Maron

Symmetry and Geometry in Neural Representations (NeurReps) @ NeurIPS 2024

Summary

We study the extent to which pretrained Graph Neural Networks can be applied across datasets, requiring feature-agnostic approaches. We build upon a purely structural pretraining approach and propose an extension to capture feature information while still being feature-agnostic. Our preliminary results indicate that embeddings from pretrained models improve generalization only with enough downstream data points and in a degree which depends on the quantity and properties of pretraining data.

Paper

Future Graph ML

Future Directions in Foundations of Graph Machine Learning

Christopher Morris, Nadav Dym, Haggai Maron, İsmail İlkan Ceylan, Fabrizio Frasca, Ron Levie, Derek Lim, Michael Bronstein, Martin Grohe, Stefanie Jegelka

International Conference on Machine Learning (ICML) 2024

Summary

We argue that the graph machine learning community needs to shift its attention to developing a more balanced theory of graph machine learning, focusing on a more thorough understanding of the interplay of expressive power, generalization, and optimization.

Paper

Spectral Expressive Power

On the Expressive Power of Spectral Invariant Graph Neural Networks

Bohang Zhang, Lingxiao Zhao, Haggai Maron

International Conference on Machine Learning (ICML) 2024

Summary

We introduce a novel message-passing framework for designing spectral invariant GNNs, called Eigenspace Projection GNN (EPNN). Our comprehensive analysis shows that EPNN essentially unifies all prior spectral invariant architectures, and they are either strictly less expressive or equivalent to EPNN. We present a surprising result that EPNN itself is bounded by a recently proposed class of Subgraph GNNs, implying that all spectral invariant architectures are strictly less expressive than 3-WL.

Paper

Deep-Align

Equivariant Deep Weight Space Alignment

Aviv Navon, Aviv Shamsian, Ethan Fetaya, Gal Chechik, Nadav Dym, Haggai Maron

International Conference on Machine Learning (ICML) 2024

Summary

We propose Deep-Align, a novel framework aimed at learning to solve the weight alignment problem. We demonstrate that weight alignment adheres to two fundamental symmetries and propose a deep architecture that respects these symmetries. Our framework does not require any labeled data. Our experimental results indicate that a feed-forward pass with Deep-Align produces better or equivalent alignments compared to those produced by current optimization algorithms.

Paper GitHub

Subgraphormer

Subgraphormer: Unifying Subgraph GNNs and Graph Transformers via Graph Products

Guy Bar-Shalom, Beatrice Bevilacqua, Haggai Maron

International Conference on Machine Learning (ICML) 2024

Summary

We propose Subgraphormer, an architecture that integrates Subgraph GNNs and Graph Transformers, combining enhanced expressive power and message-passing mechanisms from Subgraph GNNs with attention and positional encodings from Graph Transformers. Our method is based on a new connection between Subgraph GNNs and product graphs, suggesting that Subgraph GNNs can be formulated as MPNNs operating on a product of the graph with itself.

Paper

Weight Space Augmentation

Improved Generalization of Weight Space Networks via Augmentations

Aviv Shamsian, Aviv Navon, David W. Zhang, Yan Zhang, Ethan Fetaya, Gal Chechik, Haggai Maron

International Conference on Machine Learning (ICML) 2024

Oral Presentation at NeurReps Workshop @ NeurIPS 2023

Summary

We empirically analyze the reasons for overfitting in deep weight space models and find that a key reason is the lack of diversity in DWS datasets. We explore strategies for data augmentation in weight spaces and propose a MixUp method adapted for weight spaces. These methods improve classification performance similarly to having up to 10 times more data, and yield substantial 5–10% gains in downstream classification under contrastive learning.

Paper GitHub

Policy-Learn

Efficient Subgraph GNNs by Learning Effective Selection Policies

Beatrice Bevilacqua, Moshe Eliasof, Eli Meirom, Bruno Ribeiro, Haggai Maron

International Conference on Learning Representations (ICLR) 2024

Summary

We consider the problem of learning to select a small subset of the large set of possible subgraphs in a data-driven fashion. We prove that there are families of WL-indistinguishable graphs for which there exist efficient subgraph selection policies. We then propose Policy-Learn, that learns how to select subgraphs in an iterative manner. We prove that, unlike popular random policies and prior work, our architecture is able to learn these efficient policies.

Paper

Graph Metanetworks

Graph Metanetworks for Processing Diverse Neural Architectures

Derek Lim, Haggai Maron, Marc T. Law, Jonathan Lorraine, James Lucas

International Conference on Learning Representations (ICLR) 2024

Spotlight Presentation

Summary

We build new metanetworks – neural networks that take weights from other neural networks as input – by carefully building graphs representing the input neural networks and processing them using graph neural networks. Our approach, Graph Metanetworks (GMNs), generalizes to neural architectures where competing methods struggle, such as multi-head attention layers, normalization layers, convolutional layers, ResNet blocks, and group-equivariant linear layers. We prove that GMNs are expressive and equivariant to parameter permutation symmetries.

Paper

Sign Equivariant Networks

Expressive Sign Equivariant Networks for Spectral Geometric Learning

Derek Lim, Joshua Robinson, Stefanie Jegelka, Haggai Maron

37th Annual Conference on Neural Information Processing Systems (NeurIPS 2023)

Spotlight Presentation

Summary

We demonstrate that sign equivariance is useful for applications such as building orthogonally equivariant models and link prediction. We develop novel sign equivariant neural network architectures based on an analytic characterization of the sign equivariant polynomials and thus inherit provable expressiveness properties.

Norm-guided latent space

Norm-Guided Latent Space Exploration for Text-to-Image Generation

Dvir Samuel, Rami Ben-Ari, Nir Darshan, Haggai Maron, Gal Chechik

37th Annual Conference on Neural Information Processing Systems (NeurIPS 2023)

Summary

We make the observation that current training procedures make diffusion models biased toward inputs with a narrow range of norm values. We propose a novel method for interpolating between two seeds that defines a new non-Euclidean metric taking into account a norm-based prior on seeds. We show that our new interpolation and centroid evaluation techniques significantly enhance the generation of rare concept images, leading to state-of-the-art performance on few-shot and long-tail benchmarks.

Paper Code

DWS Nets

Equivariant Architectures for Learning in Deep Weight Spaces

Aviv Navon, Aviv Shamsian, Idan Achituve, Ethan Fetaya, Gal Chechik, Haggai Maron

International Conference on Machine Learning (ICML) 2023

Oral Presentation

Summary

We present a novel network architecture for learning in deep weight spaces. It takes as input a concatenation of weights and biases of a pre-trained MLP and processes it using a composition of layers that are equivariant to the natural permutation symmetry of the MLP's weights. We provide a full characterization of all affine equivariant and invariant layers for these symmetries and demonstrate the effectiveness of our architecture on a variety of learning tasks.

Paper Talk Code Blog

RFP

Graph Positional Encoding via Random Feature Propagation

Moshe Eliasof, Fabrizio Frasca, Beatrice Bevilacqua, Eran Treister, Gal Chechik, Haggai Maron

International Conference on Machine Learning (ICML) 2023

Summary

We propose Random Feature Propagation (RFP), a novel family of positional encoding schemes that draws a link between random features and spectral positional encoding. RFP is inspired by the power iteration method and concatenates several intermediate steps of an iterative algorithm for computing dominant eigenvectors of a propagation matrix, starting from random node features. RFP significantly outperforms both spectral PE and random features in multiple node classification and graph classification benchmarks.

Paper

Equivariant Polynomials

Equivariant Polynomials for Graph Neural Networks

Omri Puny, Derek Lim, Bobak T. Kiani, Haggai Maron, Yaron Lipman

International Conference on Machine Learning (ICML) 2023

Oral Presentation

Summary

We introduce an alternative expressive power hierarchy for GNNs based on their ability to calculate equivariant polynomials of a certain degree. We provide a full characterization of all equivariant graph polynomials by introducing a concrete basis, significantly generalizing previous results. We propose algorithmic tools for evaluating the expressiveness of GNNs using tensor contraction sequences, and enhance the expressivity of common GNN architectures by adding polynomial features or additional operations inspired by our theory.

Paper Code

SignNet

Sign and Basis Invariant Networks for Spectral Graph Representation Learning

Derek Lim*, Joshua Robinson*, Lingxiao Zhao, Tess Smidt, Suvrit Sra, Haggai Maron, Stefanie Jegelka

International Conference on Learning Representations (ICLR 2023)

Notable-top-25% (Spotlight)

Summary

We introduce SignNet and BasisNet – new neural architectures that are invariant to all requisite symmetries and hence process collections of eigenspaces in a principled manner. Our networks are universal and theoretically strong for graph representation learning – they can approximate any spectral graph convolution, compute spectral invariants beyond MPNNs, and provably simulate previously proposed graph positional encodings.

Paper Code

WL Survey

Weisfeiler and Leman go Machine Learning: The Story so far

Christopher Morris, Yaron Lipman, Haggai Maron, Bastian Rieck, Nils M. Kriege, Martin Grohe, Matthias Fey, Karsten Borgwardt

JMLR 2023

Summary

We give a comprehensive overview of the Weisfeiler-Leman algorithm's use in a machine learning setting, focusing on the supervised regime. We discuss the theoretical background, show how to use it for supervised graph- and node representation learning, discuss recent extensions, and outline the algorithm's connection to permutation-equivariant neural architectures.

Paper

SUN / Rethinking Subgraph GNNs

Understanding and Extending Subgraph GNNs by Rethinking Their Symmetries

Fabrizio Frasca*, Beatrice Bevilacqua*, Michael M. Bronstein, Haggai Maron

36th Annual Conference on Neural Information Processing Systems (NeurIPS 2022)

Oral Presentation (1.7% acceptance rate)

Summary

We study the most prominent form of subgraph methods employing node-based subgraph selection policies. A novel symmetry analysis shows that modeling the symmetries of node-based subgraph collections requires a significantly smaller symmetry group than the one adopted in previous works. We bound the expressive power of subgraph methods by 3-WL and propose SUN, a general family of message-passing layers for subgraph methods that generalizes all previous node-based Subgraph GNNs.

Paper Code Talk

Generalized Laplacian PE

Generalized Laplacian Positional Encoding for Graph Representation Learning

Sohir Maskey*, Ali Parviz*, Maximilian Thiessen, Hannes Stärk, Ylli Sadikaj, Haggai Maron

NeurIPS 2022 Workshop on Symmetry and Geometry in Neural Representations

Summary

We define a novel family of positional encoding schemes for graphs by generalizing the optimization problem that defines the Laplace embedding to more general dissimilarity functions rather than the 2-norm used in the original formulation. This family of positional encodings is instantiated by considering p-norms and can improve the expressive power of MPNNs.

Paper

Rotation Equivariant Network

A Simple and Universal Rotation Equivariant Point-Cloud Network

Ben Finkelshtein, Chaim Baskin, Haggai Maron, Nadav Dym

Workshop on Topology, Algebra, and Geometry in Learning, ICML 2022

Summary

We suggest a much simpler architecture than the equivariant Tensor Field Network, prove that it enjoys the same universality guarantees, and evaluate its performance on Modelnet40.

Paper Code

StyleGAN-NADA

StyleGAN-NADA: CLIP-Guided Domain Adaptation of Image Generators

Rinon Gal, Or Patashnik, Haggai Maron, Gal Chechik, Daniel Cohen-Or

ACM SIGGRAPH 2022

Summary

Leveraging the semantic power of large scale CLIP models, we present a text-driven method that allows shifting a generative model to new domains without collecting even a single image from those domains. Through natural language prompts and a few minutes of training, our method can adapt a generator across a multitude of domains characterized by diverse styles and shapes.

Paper Code Interactive Demo

Nash-MTL

Multi-Task Learning as a Bargaining Game

Aviv Navon, Aviv Shamsian, Idan Achituve, Haggai Maron, Kenji Kawaguchi, Gal Chechik, Ethan Fetaya

International Conference on Machine Learning (ICML) 2022

Summary

We propose viewing the gradients combination step in MTL as a bargaining game, where tasks negotiate to reach an agreement on a joint direction of parameter update. Under certain assumptions, the bargaining problem has a unique solution, known as the Nash Bargaining Solution, which we propose to use as a principled approach to multi-task learning. We describe Nash-MTL and derive theoretical guarantees for its convergence, achieving state-of-the-art results on multiple benchmarks.

Paper

Tensor Contraction RL

Optimizing Tensor Network Contraction Using Reinforcement Learning

Eli A Meirom, Haggai Maron, Shie Mannor, Gal Chechik

International Conference on Machine Learning (ICML) 2022

Summary

We propose an RL approach combined with GNNs to address the contraction ordering problem for quantum circuit simulation. We show how a carefully implemented RL-agent using a GNN as the basic policy construct can obtain significant improvements over state-of-the-art techniques in three varieties of circuits.

Paper

Federated Graph HyperNetworks

Federated Learning with Heterogeneous Architectures using Graph HyperNetworks

Or Litany, Haggai Maron, David Acuna, Jan Kautz, Gal Chechik, Sanja Fidler

Technical report, 2022

Summary

We propose a new FL framework that accommodates heterogeneous client architectures by adopting a graph hypernetwork for parameter sharing. Unlike existing solutions, our framework does not limit clients to sharing the same architecture type, makes no use of external data, and does not require clients to disclose their model architecture.

Paper

ESAN

Equivariant Subgraph Aggregation Networks

Beatrice Bevilacqua*, Fabrizio Frasca*, Derek Lim*, Balasubramaniam Srinivasan, Chen Cai, Gopinath Balamurugan, Michael M. Bronstein, Haggai Maron (*equal contribution)

International Conference on Learning Representations (ICLR) 2022

Spotlight Presentation (5% acceptance rate)

Summary

We propose ESAN, a novel framework where each graph is represented as a set of subgraphs derived by some predefined policy, and processed using a suitable equivariant architecture. We develop novel variants of the 1-WL test for graph isomorphism and prove lower bounds on the expressiveness of ESAN. We further prove that our approach increases the expressive power of both MPNNs and more expressive architectures.

Paper GitHub Blog Post Talk

Equivariant SfM

Deep Permutation Equivariant Structure from Motion

Dror Moran, Hodaya Koslowsky, Yoni Kasten, Haggai Maron, Meirav Galun, Ronen Basri

International Conference on Computer Vision (ICCV) 2021

Oral Presentation (3% acceptance rate)

Summary

We propose a neural network architecture that recovers both camera parameters and a sparse scene structure from point tracks in multiple images by minimizing an unsupervised reprojection loss. Our network architecture is designed to be equivariant to permutations of both cameras and scene points. We show that a pre-trained network can be used to reconstruct novel scenes using inexpensive fine-tuning with no loss of accuracy.

Paper Code

Secondary Vertex Finding

Secondary Vertex Finding in Jets with Neural Networks

Jonathan Shlomi, Sanmay Ganguly, Eilam Gross, Kyle Cranmer, Yaron Lipman, Hadar Serviansky, Haggai Maron, Nimrod Segol

European Physical Journal C, 2021

Summary

We use a neural network to perform vertex finding inside jets to improve jet classification performance, with a focus on b vs. c flavor tagging. We implement a novel, universal set-to-graph model, which takes into account information from all tracks in a jet to determine if pairs of tracks originated from a common vertex.

Paper

Dereverberation

Scene-Agnostic Multi-Microphone Speech Dereverberation

Yochai Yemini, Ethan Fetaya, Haggai Maron, Sharon Gannot

INTERSPEECH 2021

Summary

We present an NN architecture that is able to cope with microphone arrays on which no prior knowledge is presumed, and demonstrate its applicability on the speech dereverberation problem. Our approach harnesses the Deep Sets framework to design an architecture that enhances the reverberant log-spectrum.

Paper Code

Size Generalization

From Local Structures to Size Generalization in Graph Neural Networks

Gilad Yehudai, Ethan Fetaya, Eli Meirom, Gal Chechik, Haggai Maron

International Conference on Machine Learning (ICML) 2021

Summary

We put forward the size-generalization question and characterize important aspects theoretically and empirically. We show that even for very simple tasks, GNNs do not naturally generalize to graphs of larger size. Their generalization performance is closely related to the distribution of patterns of connectivity and features. We formalize size generalization as a domain-adaptation problem and describe two learning setups where it can be improved.

Paper

Graph Dynamics RL

Controlling Graph Dynamics with Reinforcement Learning and Graph Neural Networks

Eli A Meirom, Haggai Maron, Shie Mannor, Gal Chechik

International Conference on Machine Learning (ICML) 2021

Summary

We formulate partially-observed dynamic graph processes (e.g., epidemic spread) as a sequential decision problem and design RLGN, a novel tractable RL scheme to prioritize which nodes should be tested, using GNNs to rank the graph nodes. RLGN consistently outperforms all baselines and can increase the number of healthy people by 25% and contain the epidemic 30% more often than supervised approaches using the same resources.

Paper

Rotation Equivariant Point Clouds

On the Universality of Rotation Equivariant Point Cloud Networks

Nadav Dym, Haggai Maron

International Conference on Learning Representations (ICLR) 2021

Summary

We present the first study of the approximation power of rotation-equivariant point cloud network architectures. We derive two sufficient conditions for an equivariant architecture to have the universal approximation property, and use these conditions to show that two recently suggested models are universal and to devise two other novel universal architectures.

Paper

AuxiLearn

Auxiliary Learning by Implicit Differentiation

Aviv Navon*, Idan Achituve*, Haggai Maron, Gal Chechik**, Ethan Fetaya** (*/** equal contribution)

International Conference on Learning Representations (ICLR) 2021

Summary

We propose AuxiLearn, a novel framework based on implicit differentiation that targets both designing useful auxiliary tasks and combining them into a single coherent loss. We propose learning a network that combines all losses into a single coherent objective function that can learn non-linear interactions between auxiliary tasks. AuxiLearn consistently improves accuracy compared with competing methods on image segmentation and learning with attributes.

Paper GitHub

3D SSL Point Clouds

Self-Supervised Learning for Domain Adaptation on Point-Clouds

Idan Achituve, Haggai Maron, Gal Chechik

Winter Conference on Applications of Computer Vision (WACV), 2021

Summary

We describe the first study of SSL for domain adaptation on point-clouds. We introduce a new pretext task, Region Reconstruction, motivated by deformations encountered in sim-to-real transformation, and combine it with a training procedure motivated by the MixUp method. Evaluations on six domain adaptations across synthetic and real furniture data demonstrate large improvement over previous work.

Paper GitHub

Set2Graph

Set2Graph: Learning Graphs From Sets

Hadar Serviansky, Nimrod Segol, Jonathan Shlomi, Kyle Cranmer, Eilam Gross, Haggai Maron, Yaron Lipman

34th Annual Conference on Neural Information Processing Systems (NeurIPS 2020)

Summary

We suggest a neural network model family for learning Set2Graph functions that is both practical and of maximal expressive power (universal), that is, can approximate arbitrary continuous Set2Graph functions over compact sets. Testing our models on different machine learning tasks, including an application to particle physics, we find them favorable to existing baselines.

Paper Code

DSS

On Learning Sets of Symmetric Elements

Haggai Maron, Or Litany, Gal Chechik, Ethan Fetaya

International Conference on Machine Learning (ICML) 2020

ICML 2020 Outstanding Paper Award (Best Paper)

Summary

We present a principled approach to learning sets of general symmetric elements. We first characterize the space of linear layers that are equivariant both to element reordering and to the inherent symmetries of elements, like translation in the case of images. We show that networks composed of these layers (Deep Sets for Symmetric elements, DSS) are universal approximators of both invariant and equivariant functions, and they improve over existing set-learning architectures in experiments with images, graphs, and point-clouds.

Paper Code Video Slides Interview

GNN AMG

Learning Algebraic Multigrid Using Graph Neural Networks

Ilay Luz, Meirav Galun, Haggai Maron, Ronen Basri, Irad Yavneh

International Conference on Machine Learning (ICML) 2020

Summary

We propose a framework for learning AMG prolongation operators for linear systems with sparse symmetric positive (semi-)definite matrices. We train a single graph neural network to learn a mapping from an entire class of such matrices to prolongation operators, using an efficient unsupervised loss function. Experiments demonstrate improved convergence rates compared to classical AMG, demonstrating the potential utility of neural networks for developing sparse system solvers.

Paper Video+Slides Code

Open Problems IGNs

Open Problems: Approximation Power of Invariant Graph Networks

Haggai Maron, Heli Ben-Hamu, Yaron Lipman

NeurIPS 2019 Graph Representation Learning Workshop

Summary

We formulate several open problems aiming at characterizing the trade-off between expressive power and complexity of invariant and equivariant graph networks, building on recent work showing these models to be universal, in contrast to popular message passing models.

Paper Slides Talk (1:15:00)

PhD Thesis

Ph.D. Thesis: Deep and Convex Shape Analysis

Haggai Maron

Weizmann Institute of Science, 2019

Summary

This dissertation summarizes the main results obtained during my Ph.D. studies at the Weizmann Institute of Science under the guidance of Professor Yaron Lipman. Two fundamental problems in shape analysis were considered: (1) how to apply deep learning techniques to irregular data and (2) how to compute meaningful maps between shapes. My work resulted in several novel methods for applying deep learning to surfaces, point clouds, graphs and hyper-graphs as well as new efficient techniques to solve relaxations of well-known matching problems.

Thesis

SIGGRAPH DC

Deep and Convex Shape Analysis

Haggai Maron

SIGGRAPH 2019 Doctoral Consortium

Summary

I review the main results obtained during my Ph.D. studies, considering two fundamental problems in shape analysis: (1) how to apply deep learning techniques to geometric objects and (2) how to compute meaningful maps between shapes. My work resulted in novel methods for applying deep learning to surfaces, point clouds, and hyper-graphs as well as new efficient techniques to solve relaxations of well-known matching problems.

Paper Poster

Provably Powerful GNNs

Provably Powerful Graph Networks

Haggai Maron*, Heli Ben-Hamu*, Hadar Serviansky*, Yaron Lipman (*equal contribution)

33rd Annual Conference on Neural Information Processing Systems (NeurIPS 2019)

Summary

Building upon k-order invariant and equivariant graph neural networks, we present two results: First, we show that k-order networks can distinguish between non-isomorphic graphs as well as the k-WL tests, which are provably stronger than the 1-WL test for k > 2. Second, we show that a reduced 2-order network containing just scaled identity operator augmented with a single quadratic operation (matrix multiplication) has provable 3-WL expressive power. This model interleaves applications of MLP applied to the feature dimension and matrix multiplication, achieving state-of-the-art results on graph classification and regression tasks.

Arxiv GitHub (TF) GitHub (PyTorch) Blog Post Poster

Neural Level Sets

Controlling Neural Level Sets

Matan Atzmon, Niv Haim, Lior Yariv, Ofer Israelov, Haggai Maron, Yaron Lipman

33rd Annual Conference on Neural Information Processing Systems (NeurIPS 2019)

Summary

We present a simple and scalable approach to directly control level sets of a deep neural network. Our method consists of two parts: (i) sampling of the neural level sets, and (ii) relating the samples' positions to the network parameters via a sample network. We test our method on training networks robust to adversarial attacks, improving generalization to unseen data, and curve and surface reconstruction from point clouds.

Arxiv Code Poster

General Covers

Surface Networks via General Covers

Niv Haim*, Nimrod Segol*, Heli Ben-Hamu, Haggai Maron, Yaron Lipman (*equal contribution)

International Conference on Computer Vision (ICCV) 2019

Summary

We develop a novel surface-to-image representation based on a covering map from the image domain to the surface. The map wraps around the surface several times, providing a low distortion coverage of all surface parts in a single image. We use this representation to apply standard CNN models to semantic shape segmentation and shape retrieval, achieving state-of-the-art results.

Arxiv

Universality of Invariant Networks

On the Universality of Invariant Networks

Haggai Maron, Ethan Fetaya, Nimrod Segol, Yaron Lipman

International Conference on Machine Learning (ICML) 2019

Summary

We consider the fundamental question: Can G-invariant networks approximate any continuous invariant function? We tackle the general case where G ≤ S_n acts on R^n by permuting coordinates. We present two main results: First, G-invariant networks are universal if high-order tensors are allowed. Second, there are groups G for which higher-order tensors are unavoidable for obtaining universality.

Arxiv Poster Video Blog Post

IGN

Invariant and Equivariant Graph Networks

Haggai Maron, Heli Ben-Hamu, Nadav Shamir and Yaron Lipman

International Conference on Learning Representations (ICLR) 2019

Summary

We provide a characterization of all permutation invariant and equivariant linear layers for (hyper-)graph data, and show that their dimension, in case of edge-value graph data, is 2 and 15, respectively. More generally, for graph data defined on k-tuples of nodes, the dimension is the k-th and 2k-th Bell numbers. Orthogonal bases for the layers are computed. The constant number of basis elements and their characteristics allow successfully applying the networks to different size graphs.

Arxiv GitHub Poster Blog Post

Sinkhorn Lifted Assignment

Sinkhorn Algorithm for Lifted Assignment Problems

Yam Kushinsky, Haggai Maron, Nadav Dym and Yaron Lipman

SIAM Journal on Imaging Sciences, 2019

Summary

We generalize the Sinkhorn projection algorithm to higher dimensional polytopes originated from well-known lifted linear program relaxations of the Markov Random Field (MRF) energy minimization problem and the Quadratic Assignment Problem (QAP). We derive a closed-form projection on one-sided local polytopes and devise a provably convergent algorithm to solve regularized linear program relaxations of MRF and QAP.

Arxiv

Concave Graph Matching

(Probably) Concave Graph Matching

Haggai Maron and Yaron Lipman

32nd Annual Conference on Neural Information Processing Systems (NeurIPS 2018)

Spotlight Presentation (3.5% acceptance rate)

Summary

We analyze and generalize the idea of concave relaxations for graph matching. We introduce the concepts of conditionally concave and probably conditionally concave energies on polytopes and show that they encapsulate many instances of the graph matching problem, including matching Euclidean graphs and graphs on surfaces. We prove that local minima of probably conditionally concave energies on general matching polytopes are with high probability extreme points.

Arxiv Poster Short Video GitHub

Multi-chart Surface Modeling

Multi-Chart Generative Surface Modeling

Heli Ben-Hamu, Haggai Maron, Itay Kezurer, Gal Avineri and Yaron Lipman

ACM SIGGRAPH Asia 2018

Summary

A new image-like tensor data representation for genus-zero 3D shapes is devised, based on multiple parameterizations (charts) focusing on different parts of the shape. The new tensor representation is used as input to Generative Adversarial Networks for 3D shape generation, learning the shape distribution and enabling generation of novel shapes, interpolation, and exploration of the generated shape space.

Arxiv GitHub

PCNN

Point Convolutional Neural Networks by Extension Operators

Matan Atzmon*, Haggai Maron* and Yaron Lipman (*equal contribution)

ACM SIGGRAPH 2018

Summary

We present Point Convolutional Neural Networks (PCNN): a novel framework for applying convolutional neural networks to point clouds. The framework consists of extension and restriction operators, mapping point cloud functions to volumetric functions and vice-versa. PCNN is computationally efficient, invariant to the order of points, robust to different samplings and varying densities, and translation invariant. Evaluation on three central point cloud learning benchmarks convincingly outperforms competing methods.

Arxiv GitHub Slides

DS++

DS++: A Flexible, Scalable and Provably Tight Relaxation for Matching Problems

Nadav Dym*, Haggai Maron* and Yaron Lipman (*equal contribution)

ACM SIGGRAPH Asia 2017

Summary

We present a convex quadratic programming relaxation which is provably stronger than both DS and spectral relaxations, with the same scalability as the DS relaxation. The derivation naturally suggests a projection method for achieving meaningful integer solutions. Our method can be extended to optimization over doubly stochastic matrices, partial or injective matching, and problems with additional linear constraints.

Arxiv GitHub Slides

Toric CNN

Convolutional Neural Networks on Surfaces via Seamless Toric Covers

Haggai Maron, Meirav Galun, Noam Aigerman, Miri Trope, Nadav Dym, Ersin Yumer, Vladimir G. Kim and Yaron Lipman

ACM SIGGRAPH 2017

Summary

We present a method for applying deep learning to sphere-type shapes using a global seamless parameterization to a planar flat-torus, for which the convolution operator is well defined. This allows standard deep learning frameworks to be readily applied for learning semantic, high-level properties of the shape. We demonstrate success in human body segmentation and automatic landmark detection on anatomical surfaces.

Paper GitHub Data Slides Video

Point Registration

Point Registration via Efficient Convex Relaxation

Haggai Maron, Nadav Dym, Itay Kezurer, Shahar Kovalsky and Yaron Lipman

ACM SIGGRAPH 2016

Summary

We introduce a novel and efficient convex SDP relaxation for the Procrustes Matching (PM) problem. We prove that for generic isometric or almost-isometric problems the algorithm returns a correct global solution. We show our algorithm gives state-of-the-art results on popular shape matching datasets and for anatomical classification of shapes.

Paper GitHub Slides Video

Light Sensitive Display

Passive Light and Viewpoint Sensitive Display of 3D Content

Anat Levin, Haggai Maron and Michal Yarom

International Conference on Computational Photography (ICCP) 2016

Summary

We present a 3D light-sensitive display capable of presenting simple opaque 3D surfaces without self-occlusions, while reproducing both viewpoint-sensitive depth parallax and illumination-sensitive variations such as shadows and highlights. Our display is passive and uses two layers of Spatial Light Modulators whose micron-sized elements allow us to digitally simulate thin optical surfaces with flexible shapes.

Paper Paper+Supplementary