Search Results for author: Benjamin Devillers

Found 4 papers, 2 papers with code

When does CLIP generalize better than unimodal models? When judging human-centric concepts

no code implementations RepL4NLP (ACL) 2022 Romain Bielawski, Benjamin Devillers, Tim Van De Cruys, Rufin VanRullen

We compare CLIP’s visual stream against two visually trained networks and CLIP’s textual stream against two linguistically trained networks, as well as multimodal combinations of these networks.

Classification Contrastive Learning +3

Zero-shot cross-modal transfer of Reinforcement Learning policies through a Global Workspace

no code implementations7 Mar 2024 Léopold Maytié, Benjamin Devillers, Alexandre Arnold, Rufin VanRullen

First, we train a 'Global Workspace' to exploit information collected about the environment via two input modalities (a visual input, or an attribute vector representing the state of the agent and/or its environment).

Attribute Contrastive Learning +1

Semi-supervised Multimodal Representation Learning through a Global Workspace

1 code implementation27 Jun 2023 Benjamin Devillers, Léopold Maytié, Rufin VanRullen

Recent deep learning models can efficiently combine inputs from different modalities (e. g., images and text) and learn to align their latent representations, or to translate signals from one domain to another (as in image captioning, or text-to-image generation).

Image Captioning Representation Learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.