TripletCLIP: Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives

dc.contributor.authorPatel, Maitreya
dc.contributor.authorKusumba, Abhiram
dc.contributor.authorCheng, Sheng
dc.contributor.authorKim, Changhoon
dc.contributor.authorGokhale, Tejas
dc.contributor.authorBaral, Chitta
dc.contributor.authorYang, Yezhou
dc.date.accessioned2024-12-11T17:02:28Z
dc.date.available2024-12-11T17:02:28Z
dc.date.issued2024-11-04
dc.description.abstractContrastive Language-Image Pretraining (CLIP) models maximize the mutual information between text and visual modalities to learn representations. This makes the nature of the training data a significant factor in the efficacy of CLIP for downstream tasks. However, the lack of compositional diversity in contemporary image-text datasets limits the compositional reasoning ability of CLIP. We show that generating ``hard'' negative captions via in-context learning and synthesizing corresponding negative images with text-to-image generators offers a solution. We introduce a novel contrastive pre-training strategy that leverages these hard negative captions and images in an alternating fashion to train CLIP. We demonstrate that our method, named TripletCLIP, when applied to existing datasets such as CC3M and CC12M, enhances the compositional capabilities of CLIP, resulting in an absolute improvement of over 9% on the SugarCrepe benchmark on an equal computational budget, as well as improvements in zero-shot image classification and image retrieval. Our code, models, and data are available at: https://tripletclip.github.io
dc.description.sponsorshipThis work was supported by NSF RI grants #1750082, #2132724, and CPS grant #2038666. We thank the Research Computing (RC) at Arizona State University (ASU) for providing computing resources. The views and opinions of the authors expressed herein do not necessarily state or reflect those of the funding agencies and employers.
dc.description.urihttp://arxiv.org/abs/2411.02545
dc.format.extent24 pages
dc.genrejournal articles
dc.genrepostprints
dc.identifierdoi:10.13016/m2ksck-meo0
dc.identifier.urihttps://doi.org/10.48550/arXiv.2411.02545
dc.identifier.urihttp://hdl.handle.net/11603/37072
dc.language.isoen_US
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department
dc.relation.ispartofUMBC Faculty Collection
dc.rightsAttribution 4.0 International
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subjectComputer Science - Computation and Language
dc.subjectComputer Science - Computer Vision and Pattern Recognition
dc.titleTripletCLIP: Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-5593-2804

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
2411.02545v1.pdf
Size:
2.25 MB
Format:
Adobe Portable Document Format