Skip to yearly menu bar Skip to main content


Poster

Role of data structure in learning: compositionality vs stability to diffeomorphism

Umberto M. Tomasini · Matthieu Wyart


Abstract:

Understanding what makes high-dimensional data learnable is a fundamental question in machine learning. On the one hand, it is believed that the success of deep learning lies in its ability to build a hierarchy of representations that become increasingly more abstract with depth, going from simple features like edges to more complex concepts. On the other hand, learning to be insensitive to invariances of the task, such as smooth transformations for image datasets, has been argued to be important for deep networks and it strongly correlates with their performance. In this work, we aim to explain this observation and unify these two viewpoints. We show that by introducing sparsity to generative hierarchical models of data, the task acquires insensitivity to transformations that are discrete versions of diffeomorphisms. In particular, we introduce the Sparse Random Hierarchy Model (SRHM), where we show that a hierarchical representation is learnt precisely when such insensitivity is achieved, thereby explaining the strong correlation between the latter and performance. Moreover, we quantify how the sample complexity of CNNs learning the SRHM depends on both the sparsity and compositionality of the task.

Live content is unavailable. Log in and register to view live content