3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
Autor: | Jun Ota, Hamdi Sahloul, Shouhei Shirafuji |
---|---|
Rok vydání: | 2019 |
Předmět: |
Computer science
ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION denoising and interpolation 02 engineering and technology Classification of discontinuities 3D points projection lcsh:Chemical technology Biochemistry Article Analytical Chemistry local image feature embedding 0202 electrical engineering electronic engineering information engineering lcsh:TP1-1185 Computer vision Electrical and Electronic Engineering Invariant (mathematics) wide baseline matching Instrumentation Pose viewpoint invariance business.industry Detector 020207 software engineering Invariant (physics) nonparametric spherical k-means Atomic and Molecular Physics and Optics out-of-plane rotations Feature (computer vision) Computer Science::Computer Vision and Pattern Recognition Embedding RGB color model 6D pose estimation 020201 artificial intelligence & image processing Artificial intelligence Affine transformation business |
Zdroj: | Sensors, Vol 19, Iss 2, p 291 (2019) Sensors Volume 19 Issue 2 Sensors (Basel, Switzerland) |
ISSN: | 1424-8220 |
DOI: | 10.3390/s19020291 |
Popis: | Local image features are invariant to in-plane rotations and robust to minor viewpoint changes. However, the current detectors and descriptors for local image features fail to accommodate out-of-plane rotations larger than 25 ∘ &ndash 30 ∘ . Invariance to such viewpoint changes is essential for numerous applications, including wide baseline matching, 6D pose estimation, and object reconstruction. In this study, we present a general embedding that wraps a detector/descriptor pair in order to increase viewpoint invariance by exploiting input depth maps. The proposed embedding locates smooth surfaces within the input RGB-D images and projects them into a viewpoint invariant representation, enabling the detection and description of more viewpoint invariant features. Our embedding can be utilized with different combinations of descriptor/detector pairs, according to the desired application. Using synthetic and real-world objects, we evaluated the viewpoint invariance of various detectors and descriptors, for both standalone and embedded approaches. While standalone local image features fail to accommodate average viewpoint changes beyond 33 . 3 ∘ , our proposed embedding boosted the viewpoint invariance to different levels, depending on the scene geometry. Objects with distinct surface discontinuities were on average invariant up to 52 . 8 ∘ , and the overall average for all evaluated datasets was 45 . 4 ∘ . Similarly, out of a total of 140 combinations involving 20 local image features and various objects with distinct surface discontinuities, only a single standalone local image feature exceeded the goal of 60 ∘ viewpoint difference in just two combinations, as compared with 19 different local image features succeeding in 73 combinations when wrapped in the proposed embedding. Furthermore, the proposed approach operates robustly in the presence of input depth noise, even that of low-cost commodity depth sensors, and well beyond. |
Databáze: | OpenAIRE |
Externí odkaz: | |
Nepřihlášeným uživatelům se plný text nezobrazuje | K zobrazení výsledku je třeba se přihlásit. |