Zobrazeno 1 - 10
of 13
pro vyhledávání: '"Ruijie Tao"'
Autor:
Kaiwen Xue, Ruijie Tao, Qi Wu, Lei Zhang, Zhongyang Sun, Xing Yu, Jia Meng, Nirong Bao, Jianning Zhao
Publikováno v:
Frontiers in Cellular and Infection Microbiology, Vol 11 (2022)
The human gut microbiota has been proven to have great effects on the regulation of bone health. However, the association between gut microbiota and particle-induced osteolysis, which is the primary cause of aseptic loosening, is still unknown. In th
Externí odkaz:
https://doaj.org/article/a799ef78ecc04099bc4a2c4804407771
Publikováno v:
Lab Animal. 51:11-21
Drug delivery by lateral tail-vein injection in mice is widely used in preclinical research, but the technique is laborious to perform because the tail vein is hardly visible and too small to be cannulated. Misinjections of test components can lead t
We study a novel neural architecture and its training strategies of speaker encoder for speaker recognition without using any identity labels. The speaker encoder is trained to extract a fixed-size speaker embedding from a spoken utterance of various
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::3c67e9debc1d8a57af8de0676c82aba2
http://arxiv.org/abs/2210.15385
http://arxiv.org/abs/2210.15385
Autor:
Kaiwen, Xue, Ruijie, Tao, Qi, Wu, Lei, Zhang, Zhongyang, Sun, Xing, Yu, Jia, Meng, Nirong, Bao, Jianning, Zhao
Publikováno v:
Frontiers in Cellular and Infection Microbiology
The human gut microbiota has been proven to have great effects on the regulation of bone health. However, the association between gut microbiota and particle-induced osteolysis, which is the primary cause of aseptic loosening, is still unknown. In th
In self-supervised learning for speaker recognition, pseudo labels are useful as the supervision signals. It is a known fact that a speaker recognition model doesn't always benefit from pseudo labels due to their unreliability. In this work, we obser
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::c1ee385922788a4f6066d090d6e7a435
http://arxiv.org/abs/2110.03869
http://arxiv.org/abs/2110.03869
A speaker extraction algorithm seeks to extract the speech of a target speaker from a multi-talker speech mixture when given a cue that represents the target speaker, such as a pre-enrolled speech utterance, or an accompanying video track. Visual cue
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::53f0c617e2ffe81e6c435fc2aff9697c
http://arxiv.org/abs/2106.07150
http://arxiv.org/abs/2106.07150
Autor:
Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, Miguel Martin, Tushar Nagarajan, Ilija Radosavovic, Santhosh Kumar Ramakrishnan, Fiona Ryan, Jayant Sharma, Michael Wray, Mengmeng Xu, Eric Zhongcong Xu, Chen Zhao, Siddhant Bansal, Dhruv Batra, Vincent Cartillier, Sean Crane, Tien Do, Morrie Doulaty, Akshay Erapalli, Christoph Feichtenhofer, Adriano Fragomeni, Qichen Fu, Abrham Gebreselasie, Cristina Gonzalez, James Hillis, Xuhua Huang, Yifei Huang, Wenqi Jia, Weslie Khoo, Jachym Kolar, Satwik Kottur, Anurag Kumar, Federico Landini, Chao Li, Yanghao Li, Zhenqiang Li, Karttikeya Mangalam, Raghava Modhugu, Jonathan Munro, Tullie Murrell, Takumi Nishiyasu, Will Price, Paola Ruiz Puentes, Merey Ramazanova, Leda Sari, Kiran Somasundaram, Audrey Southerland, Yusuke Sugano, Ruijie Tao, Minh Vo, Yuchen Wang, Xindi Wu, Takuma Yagi, Ziwei Zhao, Yunyi Zhu, Pablo Arbelaez, David Crandall, Dima Damen, Giovanni Maria Farinella, Christian Fuegen, Bernard Ghanem, Vamsi Krishna Ithapu, C. V. Jawahar, Hanbyul Joo, Kris Kitani, Haizhou Li, Richard Newcombe, Aude Oliva, Hyun Soo Park, James M. Rehg, Yoichi Sato, Jianbo Shi, Mike Zheng Shou, Antonio Torralba, Lorenzo Torresani, Mingfei Yan, Jitendra Malik
We introduce Ego4D, a massive-scale egocentric video dataset and benchmark suite. It offers 3,670 hours of daily-life activity video spanning hundreds of scenarios (household, outdoor, workplace, leisure, etc.) captured by 931 unique camera wearers f
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::1501461e84d5acdefffd137ee9ba374d
Publikováno v:
ICASSP
Speaker extraction algorithm relies on the speech sample from the target speaker as the reference point to focus its attention. Such a reference speech is typically pre-recorded. On the other hand, the temporal synchronization between speech and lip
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::f745a12dfc999ceb73baf63dc2166519
http://arxiv.org/abs/2010.07775
http://arxiv.org/abs/2010.07775
Publikováno v:
INTERSPEECH
Audio-visual speaker recognition is one of the tasks in the recent 2019 NIST speaker recognition evaluation (SRE). Studies in neuroscience and computer science all point to the fact that vision and auditory neural signals interact in the cognitive pr
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::9ab0976f663c7a921fc9846c06126ea4
http://arxiv.org/abs/2008.03894
http://arxiv.org/abs/2008.03894
Publikováno v:
ACM Transactions on Multimedia Computing, Communications & Applications; Jul2024, Vol. 20 Issue 7, p1-29, 29p