Zobrazeno 1 - 2
of 2
pro vyhledávání: '"Eric Zhongcong Xu"'
Autor:
Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, Miguel Martin, Tushar Nagarajan, Ilija Radosavovic, Santhosh Kumar Ramakrishnan, Fiona Ryan, Jayant Sharma, Michael Wray, Mengmeng Xu, Eric Zhongcong Xu, Chen Zhao, Siddhant Bansal, Dhruv Batra, Vincent Cartillier, Sean Crane, Tien Do, Morrie Doulaty, Akshay Erapalli, Christoph Feichtenhofer, Adriano Fragomeni, Qichen Fu, Abrham Gebreselasie, Cristina Gonzalez, James Hillis, Xuhua Huang, Yifei Huang, Wenqi Jia, Weslie Khoo, Jachym Kolar, Satwik Kottur, Anurag Kumar, Federico Landini, Chao Li, Yanghao Li, Zhenqiang Li, Karttikeya Mangalam, Raghava Modhugu, Jonathan Munro, Tullie Murrell, Takumi Nishiyasu, Will Price, Paola Ruiz Puentes, Merey Ramazanova, Leda Sari, Kiran Somasundaram, Audrey Southerland, Yusuke Sugano, Ruijie Tao, Minh Vo, Yuchen Wang, Xindi Wu, Takuma Yagi, Ziwei Zhao, Yunyi Zhu, Pablo Arbelaez, David Crandall, Dima Damen, Giovanni Maria Farinella, Christian Fuegen, Bernard Ghanem, Vamsi Krishna Ithapu, C. V. Jawahar, Hanbyul Joo, Kris Kitani, Haizhou Li, Richard Newcombe, Aude Oliva, Hyun Soo Park, James M. Rehg, Yoichi Sato, Jianbo Shi, Mike Zheng Shou, Antonio Torralba, Lorenzo Torresani, Mingfei Yan, Jitendra Malik
We introduce Ego4D, a massive-scale egocentric video dataset and benchmark suite. It offers 3,670 hours of daily-life activity video spanning hundreds of scenarios (household, outdoor, workplace, leisure, etc.) captured by 931 unique camera wearers f
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::1501461e84d5acdefffd137ee9ba374d
Autor:
Kevin Qinghong Lin, Alex Jinpeng Wang, Mattia Soldan, Michael Wray, Rui Yan, Eric Zhongcong Xu, Difei Gao, Dima Damen, Bernard Ghanem, Wei Liu, Mike Zheng Shou
Publikováno v:
University of Bristol-PURE
Lin, K Q, Wang, A J, Soldan, M, Wray, M, Yan, R, Xu, E Z, Gao, D, Damen, D, Ghanem, B, Liu, W & Shou, M Z 2022, ' Egocentric Video-Language Pretraining ', Paper presented at Neural Information Processing Systems (NeurIPS), 6/12/20-12/12/20 .
Lin, K Q, Wang, A J, Soldan, M, Wray, M, Yan, R, Xu, E Z, Gao, D, Damen, D, Ghanem, B, Liu, W & Shou, M Z 2022, ' Egocentric Video-Language Pretraining ', Paper presented at Neural Information Processing Systems (NeurIPS), 6/12/20-12/12/20 .
Video-Language Pretraining (VLP), which aims to learn transferable representation to advance a wide range of video-text downstream tasks, has recently received increasing attention. Best performing works rely on large-scale, 3rd-person video- text da
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=dedup_wf_001::07c33b57554daf50f41f4a928623fb78
https://research-information.bris.ac.uk/en/publications/8739b625-0e5c-4e5e-9e14-de7b1f1aa12e
https://research-information.bris.ac.uk/en/publications/8739b625-0e5c-4e5e-9e14-de7b1f1aa12e