Leveraging non-Euclidean spaces for learning on structured data

Publié le
Equipe
Date de début de thèse (si connue)
automne 2023
Lieu
Vannes
Unité de recherche
IRISA - UMR 6074
Description du sujet de la thèse

Context. Most of machine learning algorithms require input data to be embedded in an Euclidean space,
that is to say represented in a vectorial form as a set of floats. In many domains though, the data (e.g.
hierarchical, graph or cyclical data) can be decomposed into a set of entities that possess an intricate
internal structure usually in the form of one or more relations between them. There is thus a need to
embed those structured data into meaningful representations and to define a similarity between them.
Nevertheless, those data do not possess a Euclidean structure but rather a hierarchical or cyclical struc-
ture and, in that case, they cannot be embedded in a Euclidean space with low distortion [1, 2]. In
the recent years, several works have shown the superiority of non-Euclidean representations such as the
Riemannian geometry, whose choice determines how the relations between points are quantified. In this
context, hyperbolic spaces has shown to be well suited to cope with hierarchical data, with state-of-the-art
performances in graph [3] or word embedding [4] tasks, image segmentation [5] or few-shot learning [6].
On the other hand, spherical spaces successfully deal with the face recognition problem [7] or with points
on lying on the surface of the Earth [8]. Those Riemannian spaces are of constant positive curvature
(spherical) or negative curvature (hyperbolic), in comparison to the Euclidean space which is of zero
curvature. From a field with marginal interest in the ML community about 5 years ago (Google Scholar
indicates 56 results when searching for “Hyperbolic spaces”+”Machine Learning” in 2017), it is now a
booming research field (with almost 500 results for 2022).
Going beyond spaces of constant curvature, product spaces [9] combine several copies of Euclidean, hy-
perbolic and spherical spaces, with the assumption that the data have complicated structure, aiming at
getting the best of all worlds. While these spaces have shown promising results, the optimal combination
of the spaces is still chosen by trial-errors, which is not acceptable in large scale applications.


Scientific objectives and expected achievements. The objective of the PhD is to fully take advantage
of the Riemannian geometry to perform learning on structured data. While embedding data into an
appropriate space leads to enhanced learning performances, most of the existing works have focused on
designing dedicated architectures [10] or dedicated metrics [11], synchronized on Euclidean developments.
Very few of them focused on specific loss functions that fully take advantage of the geometry at hand:
as an example, standard losses such as the cross-entropy loss are used in structured spaces, with little
modifications w.r.t. the space (i.e. angular soft max [7] for spherical spaces or hierarchical soft max
[12] for hyperbolic spaces). They very often ignore some available explicit prior knowledge such as class
text description, labels, or class hierarchy. The intrinsic properties of the spaces such as permutation
or rotational invariances are not taken into account either. In addition, when comparing different views
of the same data (e.g. text/image or images taken from different captors in remote sensing), embedded
in different geometries, the comparison of those incomparable spaces is challenging. The question of a dedicated loss for product spaces is still an open problem.
Those challenging questions will be first tackled by considering Gromov-Wasserstein [13]-like losses, which
have been shown to be an effective metric for comparing graph-like data in Euclidean Spaces. In partic-
ular, it allows comparing data living in different spaces, and retains some important invariances such as
the translational or rotational invariance. It has also shown its interest in dealing with graph-structured
data [14].
From an applicative view point, remote sensing data, that are inherently hierarchically organized, multi-
modal, with spectral and correlated attributes, will serve as a playground to validate the proposed meth-
ods. The problem of few-shot learning will be specifically targeted to evaluate the ability of generalization
and the performance in low-dimensional output embeddings.


Outcomes. Outcomes of the PhD will lead to publications in the machine learning community. This
project follows developments performed during the Multiscale ANR project (ANR-18-CE23-0022, PI:
Laetitia Chapel) whose aim is to learn on multi-modal hierarchically-organized remote sensing data.
 

Previous significative works performed by the team. The PhD will build over very recent developments
performed by the Obelix team, which shows the interest of taking into account the hierarchical nature
of the data by designing a specific loss in Euclidian spaces ([15], best paper award) or the superiority
of hyperbolic embedding w.r.t. Euclidean one in a remote sensing few shot learning problem [16]. New
metrics to compare data lying in hyperbolic [17] or spherical [11] spaces have also been proposed by some
members of the team.

Bibliographie

[1] M. Gromov, “Hyperbolic groups,” in Essays in group theory. Springer, 1987, pp. 75–263.
[2] F. Sala, C. De Sa, A. Gu, and C. R ́e, “Representation tradeoffs for hyperbolic embeddings,” in International
conference on machine learning. PMLR, 2018, pp. 4460–4469.
[3] M. Nickel and D. Kiela, “Poincaré embeddings for learning hierarchical representations,” Advances in neural
information processing systems, vol. 30, 2017.
[4] A. Tifrea, G. Becigneul, and O.-E. Ganea, “Poincaré glove: Hyperbolic word embeddings,” in International
Conference on Learning Representations, 2019. [Online]. Available: https://openreview.net/forum?id=
Ske5r3AqK7
[5] M. G. Atigh, J. Schoep, E. Acar, N. van Noord, and P. Mettes, “Hyperbolic image segmentation,” in Proceedings
of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 4453–4462.

[6] V. Khrulkov, L. Mirvakhabova, E. Ustinova, I. Oseledets, and V. Lempitsky, “Hyperbolic image embeddings,”
in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 6418–
6428.
[7] W. Liu, Y. Wen, Z. Yu, M. Li, B. Raj, and L. Song, “Sphereface: Deep hypersphere embedding for face
recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp.
212–220.
[8] A. Elad, Y. Keller, and R. Kimmel, “Texture mapping via spherical multi-dimensional scaling,” in International
Conference on Scale-Space Theories in Computer Vision. Springer, 2005, pp. 443–455.
[9] A. Gu, F. Sala, B. Gunel, and C. R ́e, “Learning mixed-curvature representations in product spaces,” in
International Conference on Learning Representations, 2018.
[10] O. Ganea, G. Bécigneul, and T. Hofmann, “Hyperbolic neural networks,” Advances in neural information
processing systems, vol. 31, 2018.
[11] C. Bonet, P. Berg, N. Courty, F. Septier, L. Drumetz, and M.-T. Pham, “Spherical sliced-wasserstein,” arXiv
preprint arXiv:2206.08780, 2022.
[12] L. Bertinetto, R. Mueller, K. Tertikas, S. Samangooei, and N. A. Lord, “Making better mistakes: Leveraging
class hierarchies with deep networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, 2020, pp. 12 506–12 515.
[13] F. Mémoli, “The gromov–wasserstein distance: A brief overview,” Axioms, vol. 3, no. 3, pp. 335–341, 2014.
[14] T. Vayer, L. Chapel, R. Flamary, R. Tavenard, and N. Courty, “Optimal transport for structured data with
application on graphs,” ICML, 2019.
[15] M. Hamzaoui, L. Chapel, M.-T. Pham, and S. Lefèvre, “A hierarchical prototypical network for few-shot remote
sensing scene classification,” in International Conference on Pattern Recognition and Artificial Intelligence.
Springer, 2022, pp. 208–220.
[16] ——, “Hyperbolic prototypical network for few shot remote sensing scene classification,” submitted, 2022.
[17] C. Bonet, L. Chapel, L. Drumetz, and N. Courty, “Hyperbolic sliced-wasserstein via geodesic and horospherical
projections,” arXiv preprint arXiv:2211.10066, 2022

Liste des encadrants et encadrantes de thèse

Nom, Prénom
Chapel, Laetitia
Type d'encadrement
Directeur.trice de thèse
Unité de recherche
UMR 6074
Equipe
Contact·s
Nom
Chapel, Laetitia
Email
laetitia.chapel@irisa.fr
Mots-clés
Machine learning, structured data, hyperbolic, spherical and product spaces, learning with invariances.