Similarity Graph-Based Camera Tracking for Effective 3D Geometry Reconstruction with Mobile RGB-D Camera
<p>Selection of a target frame for a given source frame. While capturing a stream of RGB-D images of <math display="inline"><semantics> <mrow> <mn>320</mn> <mo>×</mo> <mn>180</mn> </mrow> </semantics></math> pixels using a Google Tango-enabled smartphone, there happened to be a jerky motion between the 129th and 130th frames, which was confirmed by the low similarity measure (<b>c</b>). The presented method found that the 247th frame is in fact the best candidate as a target frame, for which the similarity measure increased markedly (<b>e</b>). By registering the source frame <math display="inline"><semantics> <msub> <mi mathvariant="script">F</mi> <mn>130</mn> </msub> </semantics></math> against the upcoming frame <math display="inline"><semantics> <msub> <mi mathvariant="script">F</mi> <mn>247</mn> </msub> </semantics></math>, not the preceding frame <math display="inline"><semantics> <msub> <mi mathvariant="script">F</mi> <mn>129</mn> </msub> </semantics></math>, we could actually avoid a significant pose estimation error. In (<b>c</b>) and (<b>e</b>), the green and red colors indicate the valid pixels of the source frame <math display="inline"><semantics> <msub> <mi mathvariant="script">F</mi> <mn>130</mn> </msub> </semantics></math> that respectively matched and did not match some pixels of the corresponding target frame candidates. On the other hand, the blue color in the target frame candidates represents the pixels that were matched by some source frame pixels. In our experiments, we set the similarity measure parameters as follows: <math display="inline"><semantics> <mrow> <msub> <mi>n</mi> <mrow> <mi>k</mi> <mi>e</mi> <mi>r</mi> </mrow> </msub> <mo>=</mo> <mn>5</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mi>τ</mi> <mi>I</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>/</mo> <mn>255</mn> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <msub> <mi>τ</mi> <mi>D</mi> </msub> <mo>=</mo> <mn>4</mn> </mrow> </semantics></math> (mm).</p> "> Figure 2
<p>Construction of a maximum spanning tree from an input RGB-D stream. From the similarity graph that was built for the stream shown in (<b>a</b>), our method found a maximum spanning tree whose root frame is marked in thicker lines in (<b>b</b>). Note the difference in the general appearance of the camera trajectories estimated through the linear path and the maximum spanning tree, respectively.</p> "> Figure 3
<p>Extraction of a frame sequence that is more effective for camera tracking. Compare the two frame sequences that respectively reach the same frame <math display="inline"><semantics> <msub> <mi mathvariant="script">F</mi> <mn>247</mn> </msub> </semantics></math>. While the frames were captured, the camera was moved rather fast between the 241th and 247th frames as shown in (<b>a</b>), for which a naïve application of the frame-to-frame tracking technique was destined to fail. In contrast, the similarity graph method was able to suggest a frame sequence, displayed in (<b>b</b>) that enabled the same frame-to-frame pose estimation method to track the camera more accurately. In each figure, the similarity measures between the respective frame pairs are shown.</p> "> Figure 4
<p>Repair of the maximum spanning tree. The rigid transformations estimated using the initial maximum spanning tree can increase the reliability of the similarity evaluation, which in turn allows for selecting <span class="html-italic">more similar</span> pairs of frames in the tree repair process. (<b>a</b>) and (<b>b</b>) illustrate the situation where the currently visited edge <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi mathvariant="script">F</mi> <mn>346</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="script">F</mi> <mn>271</mn> </msub> <mo>)</mo> </mrow> </semantics></math> is replaced by a new one <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi mathvariant="script">F</mi> <mn>346</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="script">F</mi> <mn>312</mn> </msub> <mo>)</mo> </mrow> </semantics></math>. Then, (<b>c</b>) displays the point cloud initially produced with respect to the entire 662 frames of an input RGB-D stream. The visually annoying artifact was mainly due to the inaccurate approximations of the similarity measure, causing some irrelevant pairs of frames to be selected. When those edges with similarity measures greater than <math display="inline"><semantics> <mrow> <msub> <mi>τ</mi> <mrow> <mi>g</mi> <mi>o</mi> <mi>o</mi> <mi>d</mi> </mrow> </msub> <mo>=</mo> <mn>0.75</mn> </mrow> </semantics></math> were used for 3D reconstruction, only two frames were left in the connected component containing the root frame (<b>d</b>). This was because most edges near the root node happened to have a low similarity measure. When the tree was repaired with the repair parameter <math display="inline"><semantics> <mrow> <msub> <mi>τ</mi> <mrow> <mi>r</mi> <mi>e</mi> <mi>p</mi> <mi>a</mi> <mi>i</mi> <mi>r</mi> </mrow> </msub> <mo>=</mo> <mn>0.8</mn> </mrow> </semantics></math>, a more robust 3D reconstruction was possible where the size of the connected component grew to 56 frames (<b>e</b>).</p> "> Figure 5
<p>Component-to-component camera tracking. Given two connected components <math display="inline"><semantics> <msub> <mi mathvariant="script">C</mi> <mi>i</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi mathvariant="script">C</mi> <mi>j</mi> </msub> </semantics></math> whose root frames in the respective maximum spanning trees are <math display="inline"><semantics> <msub> <mi mathvariant="script">F</mi> <mi>i</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi mathvariant="script">F</mi> <mi>j</mi> </msub> </semantics></math>, the rigid-body motion <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>i</mi> <mi>j</mi> </mrow> </msub> </semantics></math> that will align the two components in the common space is estimated by collectively using the next best available frame pairs <math display="inline"><semantics> <mrow> <mo>(</mo> <msub> <mi mathvariant="script">F</mi> <msub> <mi>i</mi> <mi>l</mi> </msub> </msub> <mo>,</mo> <msub> <mi mathvariant="script">F</mi> <msub> <mi>j</mi> <mi>l</mi> </msub> </msub> <mo>)</mo> </mrow> </semantics></math> connecting them. Note that the two rigid transformations <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <msub> <mi>i</mi> <mi>l</mi> </msub> <mo>,</mo> <mi>i</mi> </mrow> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>j</mi> <mo>,</mo> <msub> <mi>j</mi> <mi>l</mi> </msub> </mrow> </msub> </semantics></math> can be derived from the respective trees.</p> "> Figure 6
<p>Merge of separate components. When the similarity graph method was applied to an input stream of 843 frames with <math display="inline"><semantics> <mrow> <msub> <mi>τ</mi> <mrow> <mi>g</mi> <mi>o</mi> <mi>o</mi> <mi>d</mi> </mrow> </msub> <mo>=</mo> <mn>0.75</mn> </mrow> </semantics></math>, we obtained 12 separate components. To combine them as much as possible, we first repaired each component using <math display="inline"><semantics> <mrow> <msub> <mi>τ</mi> <mrow> <mi>r</mi> <mi>e</mi> <mi>p</mi> <mi>a</mi> <mi>i</mi> <mi>r</mi> </mrow> </msub> <mo>=</mo> <mn>0.8</mn> </mrow> </semantics></math>, and built a component graph whose edges are those with the highest similarity measure between the components. Then, after finding a maximum spanning tree each of whose vertices are shown in the small figure with its frame number, we performed the component-wise camera tracking with <math display="inline"><semantics> <mrow> <msub> <mi>τ</mi> <mrow> <mi>f</mi> <mi>a</mi> <mi>i</mi> <mi>r</mi> </mrow> </msub> <mo>=</mo> <mn>0.6</mn> </mrow> </semantics></math> while traversing the tree, computing the rigid transformations between the components. As a result, we could collect all components into a common space as shown in the large figure. Note that only 319 frames out of 843 were actually used to reproduce the final geometry.</p> "> Figure 7
<p>Comparison to an improved frame-to-frame tracking method. Each pair of images compare the point clouds created by the An et al.’s method [<a href="#B31-sensors-19-04897" class="html-bibr">31</a>] (<b>left</b>) and our method (<b>right</b>), respectively. As marked in ellipses in (<b>a</b>), the previous frame-to-frame technique often suffered from drifts around planar surfaces even at a normal camera speed, whereas the presented one could automatically remove the troublesome frames from consideration for camera tracking. When the camera movement was beyond its capability of adaptive error correction, as shown in (<b>b</b>), the An et al.’s method caused significant drifts of the camera poses. In contrast, when the pairs of source and target frames were selected carefully as proposed by the presented method, the simple frame-to-frame camera tracking produced quite robust results. Limiting the use of frames via the parameter <math display="inline"><semantics> <msub> <mi>τ</mi> <mrow> <mi>g</mi> <mi>o</mi> <mi>o</mi> <mi>d</mi> </mrow> </msub> </semantics></math> had a nice effect of automatically filtering out those frames that may produce intolerable errors in the reconstructed 3D geometry. The frame number in the respective caption indicates the size of the largest connected component of the similarity graph, which is displayed here.</p> "> Figure 8
<p>Comparison to the ElasticFusion method [<a href="#B27-sensors-19-04897" class="html-bibr">27</a>]. To see how negatively the presence of very abrupt and jerky camera motions in the RGB-D stream affects the pose estimation, we also compared our method to the performance-proven ElasticFusion system. In the usual situation, the ElasticFusion system, adopting a frame-to-model tracking, successfully reconstructed the scene. When we sporadically moved the camera very irregularly so that some camera views went outside the space of accumulated models, the ElasticFusion system (left) failed to correctly estimate the camera poses for some frames despite its several mechanisms for robust camera tracking (<b>a</b>). The situation got more aggravated as more radical camera movements were involved (<b>b</b>). In contrast, our method (right) could find the appropriate sets of source and target frames, still allowing acceptable 3D reconstructions.</p> "> Figure 9
<p>Comparison to the BundleFusion method [<a href="#B22-sensors-19-04897" class="html-bibr">22</a>]. For the test dataset of 587 RGB-D images that contained very fast and abrupt camera motion, the BundleFusion method scanned the scene robustly as intended, ensuring global model consistency. As marked in ellipse in (<b>a</b>), however, it was not easy to produce precise pose alignment against some regions corresponding to intractable camera motion mainly due to insufficient information. We also observed that trying to handle such camera motions sometimes affected local consistency negatively as marked in circle in (<b>b</b>). Rather than including as many frames as possible, our method takes a different approach where only those frames that, together, would lead to stabler 3D geometry reconstruction are selected for pose estimation; (<b>c</b>) shows the result from the similarity graph technique where 337 interrelated frames, merged from eight separate connected components of the similarity graph, were used. Note that the aliases in the point cloud marked in ellipse in (<b>c</b>) were due to the intensity/depth pixel mismatches and noises often incurred by the low-end mobile sensor, which could be removed in the postprocessing stage.</p> "> Figure 10
<p>3D world modeling using the mixed-reality technology. By placing each part of the scanned geometry on the real object in a mixed-reality environment, we could effectively build 3D models for the real-world scene.</p> "> Figure 11
<p>Indoor scanning of an office space using a Google Tango-enabled smartphone. In (<b>a</b>), we directly placed each of the five components from the similarity graph method on the real objects using the described experimental mixed-reality technique. As revealed in (<b>b</b>) and (<b>c</b>), the manual positioning of the parts through the HoloLens display, whose holographic image was sometimes ambiguous depending on lighting condition, achieved a nice alignment of the point clouds. Consequently, this led to a quite satisfactory 3D reconstruction result using the rather old smartphone equipped with a low-end RGB-D sensor of resolution <math display="inline"><semantics> <mrow> <mn>320</mn> <mo>×</mo> <mn>180</mn> </mrow> </semantics></math> pixels. Note that the rigid transformation found in the global space for each component may also be used as a good initial value for further fine-tuning the relative geometric relations between the components, enhancing the 3D reconstruction quality further. An effective method for this remains to be devised.</p> "> Figure 12
<p>Progressive 3D reconstruction from live RGB-D streams. When a mobile device with limited computational capability is used to scan a scene, the proposed computation may be performed on PCs that are remotely connected through a communication network.</p> ">
Abstract
:1. Introduction
2. Previous Work
3. Preliminaries: 6-DOF Camera Pose Estimation
4. Similarity Graph-Based Camera Pose Estimation
4.1. Similarity Measure between Two Frames
- , i.e., is valid in ,
- for some threshold , and
- for some threshold .
4.2. Construction of Similarity Graph
4.3. Extraction of Maximum Spanning Tree
4.4. Pose Estimation through Tree Traversal
5. Extending the Idea of a Similarity Graph
5.1. Local Repair of a Maximum Spanning Tree
5.2. Component-Wise Camera Tracking
6. Experiments
6.1. Computational Costs
6.2. Comparison to a Frame-to-Frame Tracking Method
6.3. Comparison to the ElasticFusion Method
6.4. Comparison to the BundleFusion Method
6.5. Towards 3D World Modeling in a Mixed-Reality Environment
6.6. Towards Progressive 3D Reconstruction from a Live RGB-D Stream
7. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Zollhöfer, M.; Stotko, P.; Görlitz, A.; Theobalt, C.; Nießner, M.; Klein, R.; Kolb, A. State of the art on 3D reconstruction with RGB-D cameras. Comput. Graph. Forum 2018, 37, 625–652. [Google Scholar] [CrossRef]
- Triggs, B.; McLauchlan, P.F.; Hartley, R.I.; Fitzgibbon, A.W. Bundle adjustment—A modern synthesis. In Proceedings of the International Workshop on Vision Algorithms: Theory and Practice, Corfu, Greece, 21–22 September 1999; pp. 298–372. [Google Scholar]
- Kümmerle, R.; Grisetti, G.; Strasdat, H.; Konolige, K.; Burgard, W. G2o: A general framework for graph optimization. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3607–3613. [Google Scholar]
- Lee, G.; Fraundorfer, F.; Pollefeys, M. Robust pose-graph loop-closures with expectation-maximization. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 556–563. [Google Scholar]
- Newcombe, R.A.; Izadi, S.; Hilliges, O.; Molyneaux, D.; Kim, D.; Davison, A.J.; Kohi, P.; Shotton, J.; Hodges, S.; Fitzgibbon, A. KinectFusion: Real-time dense surface mapping and tracking. In Proceedings of the 2011 10th IEEE International Symposium on Mixed and Augmented Reality, Basel, Switzerland, 26–29 October 2011; pp. 127–136. [Google Scholar]
- Roth, H.; Vona, M. Moving volume KinectFusion. In Proceedings of the British Machine Vision Conference, Surrey, UK, 4–7 September 2012; pp. 112.1–112.11. [Google Scholar]
- Whelan, T.; McDonald, J.; Kaess, M.; Fallon, M.; Johannsson, H.; Leonard, J. Kintinuous: Spatially extended KinectFusion. In Proceedings of the RSS Workshop on RGB-D: Advanced Reasoning with Depth Cameras, Sydney, Australia, 9–10 July 2012. [Google Scholar]
- Chen, J.; Bautembach, D.; Izadi, S. Scalable real-time volumetric surface reconstruction. ACM Trans. Graph. 2013, 32. [Google Scholar] [CrossRef]
- Nießner, M.; Zollhöfer, M.; Izadi, S.; Stamminger, M. Real-time 3D reconstruction at scale using voxel hashing. ACM Trans. Graph. 2013, 32. [Google Scholar] [CrossRef]
- Steinbrücker, F.; Sturm, J.; Cremers, D. Real-time visual odometry from dense RGB-D images. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 719–722. [Google Scholar]
- Audras, C.; Comport, A.; Meilland, M.; Rives, P. Real-time dense appearance-based SLAM for RGB-D sensors. In Proceedings of the Australasian Conference on Robotics and Automation (ACRA), Melbourne, Australia, 7–9 December 2011. [Google Scholar]
- Tykkälä, T.; Audras, C.; Comport, A. Direct iterative closest point for real-time visual odometry. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 2050–2056. [Google Scholar]
- Glocker, B.; Shotton, J.; Criminisi, A.; Izadi, S. Real-time RGB-D camera relocalization via randomized ferns for keyframe encoding. IEEE Trans. Vis. Comput. Graph 2015, 21, 571–583. [Google Scholar] [CrossRef] [PubMed]
- Endres, F.; Hess, J.; Engelhard, N.; Sturm, J.; Cremers, D.; Burgard, W. An evalution of the RGB-D SLAM system. In Proceedings of the 2012 IEEE International Conference Robotics Automation (ICRA), St. Paul, MN, USA, 14–18 May 2012; pp. 1691–1696. [Google Scholar]
- Henry, P.; Krainin, M.; Herbst, E.; Ren, X.; Fox, D. RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments. Int. J. Robot. Res. 2012, 31, 647–663. [Google Scholar] [CrossRef] [Green Version]
- Stückler, J.; Behnke, S. Integrating depth and color cues for dense multi-resolution scene mapping using RGB-D cameras. In Proceedings of the 2012 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), Hamburg, Germany, 13–15 September 2012. [Google Scholar]
- Kerl, C.; Sturm, J.; Cremers, D. Dense visual SLAM for RGB-D cameras. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Tokyo, Japan, 3–7 November 2013; pp. 2100–2106. [Google Scholar]
- Maier, R.; Sturm, J.; Cremers, D. Submap-based bundle adjustment for 3D reconstruction from RGB-D data. In Proceedings of the German Conference on Pattern Recognition, Münster, Germany, 2–5 September 2014; pp. 54–65. [Google Scholar]
- Stückler, J.; Behnke, S. Multi-resolution surfel maps for efficient dense 3D modeling and tracking. J. Vis. Commun. Image Represent. 2014, 25, 137–147. [Google Scholar] [CrossRef]
- Choi, S.; Zhou, Q.; Koltun, V. Robust reconstruction of indoor scenes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5556–5565. [Google Scholar]
- Kähler, O.; Prisacariu, V.; Murray, D. Real-time large-scale dense 3D reconstruction with loop closure. In Proceedings of the 2016 European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 500–516. [Google Scholar]
- Dai, A.; Nießner, M.; Zollöfer, M.; Izadi, S.; Theobalt, C. BundleFusion: Real-time globally consistent 3D reconstruction using on-the-fly surface re-integration. ACM Trans. Graph. 2017, 36. [Google Scholar] [CrossRef]
- Cao, Y.; Kobbelt, L.; Hu, S. Real-time high-accuracy three-dimensional reconstruction with consumer RGB-D cameras. ACM Trans. Graph. 2018, 37. [Google Scholar] [CrossRef]
- Keller, M.; Lefloch, D.; Lambers, M.; Izadi, S.; Weyrich, T.; Kolb, A. Real-time 3D reconstruction in dynamic scenes using point-based fusion. In Proceedings of the 2013 International Conference on 3D Vision, Seattle, WA, USA, 29 June–1 July 2013; 1 July 2013; pp. 1–8. [Google Scholar]
- Ruhnke, M.; Kümmerle, R.; Grisetti, G.; Burgard, W. Highly accurate 3D surface models by sparse surface adjustment. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012; pp. 751–757. [Google Scholar]
- Whelan, T.; Kaess, M.; Leonard, J.J.; McDonald, J. Deformation-based loop closure for large scale dense RGB-D SLAM. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robot Systems, Tokyo, Japan, 3–7 November 2013; pp. 548–555. [Google Scholar]
- Whelan, T.; Salas-Moreno, R.; Glocker, B.; Davison, J.; Leutenegger, S. ElasticFusion: Real-time dense SLAM and light source estimation. Int. J. Robot. Res. 2016, 35, 1697–1716. [Google Scholar] [CrossRef] [Green Version]
- Mur-Artal, R.; Tardós, J.D. ORB-SLAM2: An open-source SLAM system for monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. 2017, 33, 1225–1262. [Google Scholar] [CrossRef]
- Yousif, K.; Taguchi, Y.; Ramalingam, S. MonoRGBD-SLAM: Simultaneous localization and mapping using both monocular and RGBD cameras. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 4495–4502. [Google Scholar]
- Endres, F.; Hess, J.; Sturm, J.; Cremers, D.; Burgard, W. 3-D mapping with an RGB-D camera. IEEE Trans. Robot. 2014, 30, 177–187. [Google Scholar] [CrossRef]
- An, J.; Lee, J.; Jeong, J.; Ihm, I. Tracking an RGB-D camera on mobile devices using an improved frame-to-frame pose estimation method. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision 2018 (WACV 2018), Lake Tahoe, NV, USA, 12–14 March 2018; pp. 1142–1150. [Google Scholar]
- Solomon, J. Numerical Algorithms: Methods for Computer Vision, Machine Learning, and Graphics; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
- Sedgewick, R. Algorithms in C: Part 5 Graph Algorithms, 3rd ed.; Addison-Wesley: Reading, MA, USA, 2002. [Google Scholar]
- Handa, A.; Whelan, T.; McDonald, J.; Davison, A. A benchmark for RGB-D visual odometry, 3D reconstruction and SLAM. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–5 June 2014. [Google Scholar]
- Sturm, J.; Engelhard, N.; Endres, F.; Burgard, W.; Cremers, D. A benchmark for the evaluation of RGB-D SLAM systems. In Proceedings of the International Conference on Intelligent Robot Systems, Vilamoura, Algarve, Portugal, 7–12 October 2012. [Google Scholar]
Sim-Gr | Mst | Paths | Pose-Est | Amortized | |
---|---|---|---|---|---|
200 | 0.77 | 0.0068 | 0.0022 | 5.15 (0.0258) | 0.0297 |
400 | 2.86 | 0.0332 | 0.0044 | 10.62 (0.0266) | 0.0338 |
800 | 10.83 | 0.1879 | 0.0128 | 22.59 (0.0282) | 0.0420 |
1600 | 40.52 | 1.0450 | 0.0520 | 46.50 (0.0291) | 0.0551 |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
An, J.; Lee, S.; Park, S.; Ihm, I. Similarity Graph-Based Camera Tracking for Effective 3D Geometry Reconstruction with Mobile RGB-D Camera. Sensors 2019, 19, 4897. https://doi.org/10.3390/s19224897
An J, Lee S, Park S, Ihm I. Similarity Graph-Based Camera Tracking for Effective 3D Geometry Reconstruction with Mobile RGB-D Camera. Sensors. 2019; 19(22):4897. https://doi.org/10.3390/s19224897
Chicago/Turabian StyleAn, Jaepung, Sangbeom Lee, Sanghun Park, and Insung Ihm. 2019. "Similarity Graph-Based Camera Tracking for Effective 3D Geometry Reconstruction with Mobile RGB-D Camera" Sensors 19, no. 22: 4897. https://doi.org/10.3390/s19224897
APA StyleAn, J., Lee, S., Park, S., & Ihm, I. (2019). Similarity Graph-Based Camera Tracking for Effective 3D Geometry Reconstruction with Mobile RGB-D Camera. Sensors, 19(22), 4897. https://doi.org/10.3390/s19224897