CE23 - Intelligence artificielle et science des données 2023

Multimodal deep SensoriMotor Representation learning – MeSMRise

Submission summary

There is “a fundamental misalignment between human and typical AI representations: while the former are grounded in rich sensorimotor experience, the latter are typically passive and limited to a few modalities such as vision and text” (Hay et al, 2016). We propose in this project to take inspiration from the way human babies learn to explore their environment through actions that shape their multimodal experience. Especially, the sensorimotor contingencies (SMC) theory combines coherent pieces of evidence from neuroscience, psychology, etc. of human perception and learning in a unified framework. The key claims are learning of SMCs defined as “the structure of the rules governing the sensory changes produced by various motor actions” (O'Regan et al, 2001) and active perception as the “organism’s exploration of the environment that is mediated by knowledge of SMCs” (Myin et al, 2002). Some models implementing this theory are able to learn complex concepts such as containment for instance.

Inspired by the SMC theory, the main objective of the project is to study how action can structure the multimodal representations, learned with self-supervised learning methods. This will be applied to 3D objects, perceived by vision and point cloud, and manipulated in virtual environments. By proposing a new paradigm for unsupervised representation learning with multimodal data fusion, it will participate to multiple issues of the E.2 axis. Specifically, we target the following properties:
- generalization to unknown environments and contexts
- robustness, e.g. to the orientation, background, shape ... of the object
- adaptability via the capacity of the model to autonomously find relevant information
- generality by using similar architectures and principles for all research questions

Project coordination

Mathieu LEFORT (UMR 5205 - LABORATOIRE D'INFORMATIQUE EN IMAGE ET SYSTEMES D'INFORMATION)

The author of this summary is the project coordinator, who is responsible for the content of this summary. The ANR declines any responsibility as for its contents.

Partnership

LIRIS UMR 5205 - LABORATOIRE D'INFORMATIQUE EN IMAGE ET SYSTEMES D'INFORMATION
LJK Laboratoire Jean Kuntzmann
IP INSTITUT PASCAL

Help of the ANR 511,261 euros
Beginning and duration of the scientific project: March 2024 - 54 Months

Useful links

Explorez notre base de projets financés

 

 

ANR makes available its datasets on funded projects, click here to find more.

Sign up for the latest news:
Subscribe to our newsletter