IEEE VIS 2024 Content: Wasserstein Auto-Encoders of Merge Trees (and Persistence Diagrams)

Wasserstein Auto-Encoders of Merge Trees (and Persistence Diagrams)

Mathieu Pont -

Julien Tierny -

Room: Bayshore I

2024-10-17T15:03:00ZGMT-0600Change your timezone on the schedule page
2024-10-17T15:03:00Z
Exemplar figure, described by caption below
Visual analysis of the Earthquake ensemble ((a) each ground-truth class is represented by one of its members), with our Wasserstein Auto-Encoder of Merge Trees (MT-WAE). We apply our contributions to merge tree compression ((b), right) by simply storing their coordinates in the last decoding layer of our network. We exploit the latent space of our network to generate 2D layouts of the ensemble (c). The reconstruction of user-defined locations ((c) and (d), purple) enables an interactive exploration of the latent space. MT-WAE also supports persistence correlation views (e), which reveal the persistent features which exhibit the most variability in the ensemble.
Fast forward
Keywords

Topological data analysis, ensemble data, persistence diagrams, merge trees, auto-encoders, neural networks

Abstract

This paper presents a computational framework for the Wasserstein auto-encoding of merge trees (MT-WAE), a novel extension of the classical auto-encoder neural network architecture to the Wasserstein metric space of merge trees. In contrast to traditional auto-encoders which operate on vectorized data, our formulation explicitly manipulates merge trees on their associated metric space at each layer of the network, resulting in superior accuracy and interpretability. Our novel neural network approach can be interpreted as a non-linear generalization of previous linear attempts [79] at merge tree encoding. It also trivially extends to persistence diagrams. Extensive experiments on public ensembles demonstrate the efficiency of our algorithms, with MT-WAE computations in the orders of minutes on average. We show the utility of our contributions in two applications adapted from previous work on merge tree encoding [79]. First, we apply MT-WAE to merge tree compression, by concisely representing them with their coordinates in the final layer of our auto-encoder. Second, we document an application to dimensionality reduction, by exploiting the latent space of our auto-encoder, for the visual analysis of ensemble data. We illustrate the versatility of our framework by introducing two penalty terms, to help preserve in the latent space both the Wasserstein distances between merge trees, as well as their clusters. In both applications, quantitative experiments assess the relevance of our framework. Finally, we provide a C++ implementation that can be used for reproducibility.