Zobrazeno 1 - 10
of 1 298
pro vyhledávání: '"Vedaldi, A"'
Generating high-quality 3D content from text, single images, or sparse view images remains a challenging task with broad applications. Existing methods typically employ multi-view diffusion models to synthesize multi-view images, followed by a feed-f
Externí odkaz:
http://arxiv.org/abs/2410.00890
Image-based 3D object detection is widely employed in applications such as autonomous vehicles and robotics, yet current systems struggle with generalisation due to complex problem setup and limited training data. We introduce a novel pipeline that d
Externí odkaz:
http://arxiv.org/abs/2408.12747
Autor:
Bhalgat, Yash, Tschernezki, Vadim, Laina, Iro, Henriques, João F., Vedaldi, Andrea, Zisserman, Andrew
Egocentric videos present unique challenges for 3D scene understanding due to rapid camera motion, frequent object occlusions, and limited object visibility. This paper introduces a novel approach to instance segmentation and tracking in first-person
Externí odkaz:
http://arxiv.org/abs/2408.09860
We present Puppet-Master, an interactive video generative model that can serve as a motion prior for part-level dynamics. At test time, given a single image and a sparse set of motion trajectories (i.e., drags), Puppet-Master can synthesize a video d
Externí odkaz:
http://arxiv.org/abs/2408.04631
Canonical surface mapping generalizes keypoint detection by assigning each pixel of an object to a corresponding point in a 3D template. Popularised by DensePose for the analysis of humans, authors have since attempted to apply the concept to more ca
Externí odkaz:
http://arxiv.org/abs/2407.18907
Autor:
Bensadoun, Raphael, Monnier, Tom, Kleiman, Yanir, Kokkinos, Filippos, Siddiqui, Yawar, Kariya, Mahendra, Harosh, Omri, Shapovalov, Roman, Graham, Benjamin, Garreau, Emilien, Karnewar, Animesh, Cao, Ang, Azuri, Idan, Makarov, Iurii, Le, Eric-Tuan, Toisoul, Antoine, Novotny, David, Gafni, Oran, Neverova, Natalia, Vedaldi, Andrea
We introduce Meta 3D Gen (3DGen), a new state-of-the-art, fast pipeline for text-to-3D asset generation. 3DGen offers 3D asset creation with high prompt fidelity and high-quality 3D shapes and textures in under a minute. It supports physically-based
Externí odkaz:
http://arxiv.org/abs/2407.02599
Autor:
Siddiqui, Yawar, Monnier, Tom, Kokkinos, Filippos, Kariya, Mahendra, Kleiman, Yanir, Garreau, Emilien, Gafni, Oran, Neverova, Natalia, Vedaldi, Andrea, Shapovalov, Roman, Novotny, David
We present Meta 3D AssetGen (AssetGen), a significant advancement in text-to-3D generation which produces faithful, high-quality meshes with texture and material control. Compared to works that bake shading in the 3D object's appearance, AssetGen out
Externí odkaz:
http://arxiv.org/abs/2407.02445
Autor:
Bensadoun, Raphael, Kleiman, Yanir, Azuri, Idan, Harosh, Omri, Vedaldi, Andrea, Neverova, Natalia, Gafni, Oran
The recent availability and adaptability of text-to-image models has sparked a new era in many related domains that benefit from the learned text priors as well as high-quality and fast generation capabilities, one of which is texture generation for
Externí odkaz:
http://arxiv.org/abs/2407.02430
Autor:
Szymanowicz, Stanislaw, Insafutdinov, Eldar, Zheng, Chuanxia, Campbell, Dylan, Henriques, João F., Rupprecht, Christian, Vedaldi, Andrea
In this paper, we propose Flash3D, a method for scene reconstruction and novel view synthesis from a single image which is both very generalisable and efficient. For generalisability, we start from a "foundation" model for monocular depth estimation
Externí odkaz:
http://arxiv.org/abs/2406.04343
3D scene generation has quickly become a challenging new research direction, fueled by consistent improvements of 2D generative diffusion models. Most prior work in this area generates scenes by iteratively stitching newly generated frames with exist
Externí odkaz:
http://arxiv.org/abs/2404.19758