Less Descriptive yet Discriminative: Quantifying the Properties of Multimodal Referring Utterances via CLIP
Autor: | Takmaz, E., Pezzelle, S., Fernández, R., Chersoni, E., Hollenstein, N., Jacobs, C., Oseki, Y., Prévot, L., Santus, E. |
---|---|
Přispěvatelé: | ILLC (FNWI) |
Jazyk: | angličtina |
Rok vydání: | 2022 |
Zdroj: | Workshop on Cognitive Modeling and Computational Linguistics: CMCL 2022 : proceedings of the workshop : May 26, 2022, 36-42 STARTPAGE=36;ENDPAGE=42;TITLE=Workshop on Cognitive Modeling and Computational Linguistics |
Popis: | In this work, we use a transformer-based pre-trained multimodal model, CLIP, to shed light on the mechanisms employed by human speakers when referring to visual entities. In particular, we use CLIP to quantify the degree of descriptiveness (how well an utterance describes an image in isolation) and discriminativeness (to what extent an utterance is effective in picking out a single image among similar images) of human referring utterances within multimodal dialogues. Overall, our results show that utterances become less descriptive over time while their discriminativeness remains unchanged. Through analysis, we propose that this trend could be due to participants relying on the previous mentions in the dialogue history, as well as being able to distill the most discriminative information from the visual context. In general, our study opens up the possibility of using this and similar models to quantify patterns in human data and shed light on the underlying cognitive mechanisms. |
Databáze: | OpenAIRE |
Externí odkaz: |