Zobrazeno 1 - 10
of 10 255
pro vyhledávání: '"modal interaction"'
Despite multimodal sentiment analysis being a fertile research ground that merits further investigation, current approaches take up high annotation cost and suffer from label ambiguity, non-amicable to high-quality labeled data acquisition. Furthermo
Externí odkaz:
http://arxiv.org/abs/2412.09784
Visual Language Tracking (VLT) enhances tracking by mitigating the limitations of relying solely on the visual modality, utilizing high-level semantic information through language. This integration of the language enables more advanced human-machine
Externí odkaz:
http://arxiv.org/abs/2409.08887
The intricate nature of real-world driving environments, characterized by dynamic and diverse interactions among multiple vehicles and their possible future states, presents considerable challenges in accurately predicting the motion states of vehicl
Externí odkaz:
http://arxiv.org/abs/2409.11676
Image-text matching (ITM) is a fundamental problem in computer vision. The key issue lies in jointly learning the visual and textual representation to estimate their similarity accurately. Most existing methods focus on feature enhancement within mod
Externí odkaz:
http://arxiv.org/abs/2406.18579
Capturing complex temporal relationships between video and audio modalities is vital for Audio-Visual Emotion Recognition (AVER). However, existing methods lack attention to local details, such as facial state changes between video frames, which can
Externí odkaz:
http://arxiv.org/abs/2405.16701
Multi-modal learning that combines pathological images with genomic data has significantly enhanced the accuracy of survival prediction. Nevertheless, existing methods have not fully utilized the inherent hierarchical structure within both whole slid
Externí odkaz:
http://arxiv.org/abs/2404.08027
Multi-modal entity alignment (MMEA) aims to identify equivalent entity pairs across different multi-modal knowledge graphs (MMKGs). Existing approaches focus on how to better encode and aggregate information from different modalities. However, it is
Externí odkaz:
http://arxiv.org/abs/2404.17590
Akademický článek
Tento výsledek nelze pro nepřihlášené uživatele zobrazit.
K zobrazení výsledku je třeba se přihlásit.
K zobrazení výsledku je třeba se přihlásit.
Tour guidance in virtual museums encourages multi-modal interactions to boost user experiences, concerning engagement, immersion, and spatial awareness. Nevertheless, achieving the goal is challenging due to the complexity of comprehending diverse us
Externí odkaz:
http://arxiv.org/abs/2401.11923
Publikováno v:
Jisuanji kexue yu tansuo, Vol 18, Iss 5, Pp 1318-1327 (2024)
To address the issues of insufficient modal fusion and weak interactivity caused by semantic feature differences between different modalities in multimodal emotion analysis, a temporal multimodal sentiment analysis model for composite cross modal int
Externí odkaz:
https://doaj.org/article/17957e740a5b459c8310509ae18e9757