Don’t Buy it! Reassessing the Ad Understanding Abilities of Contrastive Multimodal Models

Open Access
Authors
Publication date 2024
Host editors
  • L.-W. Ku
  • A. Martins
  • V. Srikumar
Book title The 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024) : proceedings of the conference
Book subtitle ACL 2024 : August 11-16, 2024
ISBN (electronic)
  • 9798891760950
Event 62nd Annual Meeting of the Association for Computational Linguistics, ACL 2024
Volume | Issue number 2
Pages (from-to) 870-879
Number of pages 10
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

Image-based advertisements are complex multimodal stimuli that often contain unusual visual elements and figurative language. Previous research on automatic ad understanding has reported impressive zero-shot accuracy of contrastive vision-and-language models (VLMs) on an ad-explanation retrieval task. Here, we examine the original task setup and show that contrastive VLMs can solve it by exploiting grounding heuristics. To control for this confound, we introduce TRADE, a new evaluation test set with adversarial grounded explanations. While these explanations look implausible to humans, we show that they “fool” four different contrastive VLMs. Our findings highlight the need for an improved operationalisation of automatic ad understanding that truly evaluates VLMs’ multimodal reasoning abilities. We make our code and TRADE available at https://github.com/dmg-illc/trade.

Document type Conference contribution
Language English
Related dataset TRADE: TRuly ADversarial ad understanding Evaluation
Published at https://doi.org/10.18653/v1/2024.acl-short.77
Other links https://github.com/dmg-illc/trade https://www.scopus.com/pages/publications/85203824138
Downloads
2024.acl-short.77 (Final published version)
Permalink to this page
Back