Going Beyond Nouns With Vision & Language
Models Using Synthetic Data

Paola Cascante-Bonilla*1,2,     Khaled Shehada*2,3,     James Seale Smith2,4,     Sivan Doveh6,7,
Donghyun Kim2,7,     Rameswar Panda2,7,     Gül Varol5,     Aude Oliva2,3,
Vicente Ordonez1,     Rogerio Feris2,7,     Leonid Karlinsky2,7
1Rice University,     2MIT-IBM Watson AI Lab,     3MIT,     4Georgia Institute of Technology,
5LIGM, École des Ponts,     6Weizmann Institute of Science,     7IBM Research

Abstract

Large-scale pre-trained Vision & Language (VL) models have shown remarkable performance in many applications, enabling replacing a fixed set of supported classes with zero-shot open vocabulary reasoning over (almost arbitrary) natural language prompts. However, recent works have uncovered a fundamental weakness of these models. For example, their difficulty to understand Visual Language Concepts (VLC) that go 'beyond nouns' such as the meaning of non-object words (e.g., attributes, actions, relations, states, etc.), or difficulty in performing compositional reasoning such as understanding the significance of the order of the words in a sentence. In this work, we investigate to which extent purely synthetic data could be leveraged to teach these models to overcome such shortcomings without compromising their zero-shot capabilities. We contribute Synthetic Visual Concepts (SyViC) - a million-scale synthetic dataset and data generation codebase allowing to generate additional suitable data to improve VLC understanding and compositional reasoning of VL models. Additionally, we propose a general VL finetuning strategy for effectively leveraging SyViC towards achieving these improvements. Our extensive experiments and ablations on VL-Checklist, Winoground, and ARO benchmarks demonstrate that it is possible to adapt strong pre-trained VL models with synthetic data significantly enhancing their VLC understanding (e.g. by 9.9% on ARO and 4.3% on VL-Checklist) with under 1% drop in their zero-shot accuracy.


Overview

We finetune a large-scale pre-trained VL model using our SyViC dataset. Our proposed methodology includes domain adaptation and avoiding forgetting through parameterefficient finetuning (LoRA) and model averaging.

Summary of the entire flow, including components and choices of SyViC data synthesis pipeline (left) and the proposed effective finetuning technique (right).
Results

We evaluate our finetuned model on three benchmarks: VL-Checklist, ARO, and Winoground. The compositional reasoning evaluation includes understanding the meaning of the sentence after changing the word order, attributes, and relations of humans/objects.

Poster




BibTeX
@misc{cascantebonilla2023going,
    title={Going Beyond Nouns With Vision & Language Models Using Synthetic Data},
    author={Paola Cascante-Bonilla, Khaled Shehada, James Seale Smith, Sivan Doveh, Donghyun Kim, Rameswar Panda, Gül Varol, Aude Oliva, Vicente Ordonez, Rogerio Feris, Leonid Karlinsky},
    year={2023},
    eprint={2303.17590},
    archivePrefix={arXiv},
    primaryClass={cs.CV} }