Zobrazeno 1 - 10
of 85
pro vyhledávání: '"Malik, Sarthak"'
The ability to generate sentiment-controlled feedback in response to multimodal inputs comprising text and images addresses a critical gap in human-computer interaction. This capability allows systems to provide empathetic, accurate, and engaging res
Externí odkaz:
http://arxiv.org/abs/2402.07640
This paper proposes a multimodal emotion recognition system based on hybrid fusion that classifies the emotions depicted by speech utterances and corresponding images into discrete classes. A new interpretability technique has been developed to ident
Externí odkaz:
http://arxiv.org/abs/2208.11868
This paper proposes a multimodal emotion recognition system, VIsual Spoken Textual Additive Net (VISTANet), to classify emotions reflected by input containing image, speech, and text into discrete classes. A new interpretability technique, K-Average
Externí odkaz:
http://arxiv.org/abs/2208.11450
Autor:
Puri, Pankaj, Malik, Sarthak
Publikováno v:
In Journal of Clinical and Experimental Hepatology November-December 2023 13(6):1116-1129
Autor:
Mallick, Bipadabhanjan, Dhaka, Narendra, Sharma, Vishal, Malik, Sarthak, Sinha, Saroj K., Dutta, Usha, Gupta, Pankaj, Gulati, Ajay, Yadav, Thakur D., Gupta, Vikas, Kochhar, Rakesh
Publikováno v:
In Pancreatology January 2019 19(1):143-148
Autor:
Kochhar, Rakesh, Malik, Sarthak, Gupta, Pankaj, Reddy, Yalaka Rami, Dhaka, Narendra, Sinha, Saroj Kant, Gupta, Vikas, Noor, Mohd Talha, Mallick, Bipadabhanjan
Publikováno v:
In Gastrointestinal Endoscopy December 2018 88(6):899-908
Publikováno v:
Multimedia Tools & Applications; Mar2024, Vol. 83 Issue 10, p28373-28394, 22p
Akademický článek
Tento výsledek nelze pro nepřihlášené uživatele zobrazit.
K zobrazení výsledku je třeba se přihlásit.
K zobrazení výsledku je třeba se přihlásit.
Autor:
Mallick, Bipadabhanjan, Dhaka, Narendra, Gupta, Pankaj, Gulati, Ajay, Malik, Sarthak, Sinha, Saroj K., Yadav, Thakur D., Gupta, Vikas, Kochhar, Rakesh
Publikováno v:
In Pancreatology October 2018 18(7):727-733
This paper proposes a multimodal emotion recognition system, VIsual Spoken Textual Additive Net (VISTA Net), to classify the emotions reflected by a multimodal input containing image, speech, and text into discrete classes. A new interpretability tec
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::dc9430fcaa5a5fd55952517fe9eeec82