Share This Article:

Segmenting Salient Objects in 3D Point Clouds of Indoor Scenes Using Geodesic Distances

Abstract Full-Text HTML Download Download as PDF (Size:1160KB) PP. 102-108
DOI: 10.4236/jsip.2013.43B018    3,223 Downloads   4,574 Views   Citations

ABSTRACT

Visual attention mechanisms allow humans to extract relevant and important information from raw input percepts. Many applications in robotics and computer vision have modeled human visual attention mechanisms using a bottom-up data centric approach. In contrast, recent studies in cognitive science highlight advantages of a top-down approach to the attention mechanisms, especially in applications involving goal-directed search. In this paper, we propose a top-down approach for extracting salient objects/regions of space. The top-down methodology first isolates different objects in an unorganized point cloud, and compares each object for uniqueness. A measure of saliency using the properties of geodesic distance on the object’s surface is defined. Our method works on 3D point cloud data, and identifies salient objects of high curvature and unique silhouette. These being the most unique features of a scene, are robust to clutter, occlusions and view point changes. We provide the details of the proposed method and initial experimental results.

 

Conflicts of Interest

The authors declare no conflicts of interest.

Cite this paper

S. Bhatia and S. Chalup, "Segmenting Salient Objects in 3D Point Clouds of Indoor Scenes Using Geodesic Distances," Journal of Signal and Information Processing, Vol. 4 No. 3B, 2013, pp. 102-108. doi: 10.4236/jsip.2013.43B018.

References

[1] T. N. Vikram, M. Tscherepanow and B. Wrede, “A Saliency Map Based on Sampling an Image Into Random Rectangular Regions of Interest,” Pattern Recognition, Vol. 45, No. 9, Sep. 2013, pp. 3114-3124. doi:10.1016/j.patcog.2012.02.009
[2] S. Frintrop, E. Rome and H. I. Christensen, “Computational Visual Attention Systems and Their Cognitive Foundations,” ACM Transactions on Applied Perception, vol. 7, no. 1, Jan. 2010, pp. 1-39. doi:10.1145/1658349.1658355
[3] N. Riche, M. Mancas, B. Gosselin and T. Dutoit, “3D Saliency for Abnormal Motion Selection: The Role of the Depth Map,” in J. Crowley, B. Draper, and M. Thonnat, Eds. Computer Vision Systems, Springer Berlin/Heidel- berg, 2011, pp. 143-152. doi:10.1007/978-3-642-23968-7_15
[4] O. Akman and P. Jonker, “Computing Saliency Map from Spatial Information in Point Cloud Data,” Advanced Concepts for Intelligent Vision Systems, Vol. 6474, 2010, pp. 290-299. doi:10.1007/978-3-642-17688-3_28
[5] D. Simon, “Fast and Accurate Shape-Based Registration,” PhD thesis, Robotics Institute, Carnegie Mellon University, Pittsburg, PA, 1996.
[6] D. Cole and A. Harrison, “Using Naturally Salient Regions for SLAM with 3D Laser Data,” In Proceedings of International Conference on Robotics and Automation, Workshop on SLAM, 2005
[7] J. Stückler and S. Behnke, “Interest Point Detection in Depth Images Through Scale-Space Surface Analysis,” In 2011 IEEE International Conference on Robotics and Automation (ICRA), 2011, pp. 3568-3574. doi:10.1109/ICRA.2011.5980474
[8] C. Koch and S. Ullman, “Shifts in Selective Visual Attention: Towards the Underlying Neural Circuitry,” Human Neurobiology, Springer-Verlag, Vol. 4, No. 4, 1985, pp. 219-227.
[9] L. Itti, C. Koch and E. Niebur, “A Model of Saliency Based Visual Attention for Rapid Scene Analysis,” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 20, No. 11, 1998, pp. 1254-1259. doi:10.1109/34.730558
[10] C. E. Connor, H. E. Egeth and S. Yantis, “Visual Attention: Bottom-Up Versus Top-Down,” Current Biology, Vol. 14, No. 19, 2004, pp. R850-R852. doi:10.1016/j.cub.2004.09.041
[11] M. Begum and F. Karray, “Visual Attention for Robotic Cognition: A Survey,” IEEE Transactions on Autonomous Mental Development, Vol. 3, No. 1, 2011, pp. 92-105. doi:10.1109/TAMD.2010.2096505
[12] B. Steder and R. Rusu, “NARF: 3D Range Image Features for Object Recognition”, In Workshop on Defining and Solving Realistic Perception Problems in Personal Robotics at the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2010.
[13] R. Unnikrishnan and M. Hebert, “Multi-Scale Interest Regions From Unorganized Point Clouds”, In IEEE Computer Society Conference, Computer Vision and Pattern Recognition Workshops, 2008, pp. 1-8.
[14] A. Flint, A. Dick and A. v. d. Hengel, “Thrift: Local 3D Structure Recognition,” In Society on Digital Image Computing Techniques and Applications, 9th Biennial Conference of the Australian Pattern Recognition, 2007, pp. 182-188.
[15] E. Potapova, M. Zillich and M. Vincze, “Calculation of Attention Points Using 3D Cues,” 35th Annual Workshop of the Austrian Association for Pattern Recognition (OAGM/AAPR), May 2011.
[16] R. B. Rusu, “Semantic 3d Object Maps for Everyday Manipulation in Human Living Environments,” PhD Thesis, Computer Science Department, The University of Rochester, Oct. 2009.
[17] D. Holz, S. Holzer and R. Rusu, “Real-Time Plane Segmentation Using RGB-D Cameras,” In Proceedings of the 15th RoboCup International Symposium, Isntanbul, Turkey, 2011, pp. 306-317.
[18] Microsoft kinect sensor, http://www.xbox.com/en-au/kinect?xr=shellnav," 2013.
[19] E. Fix and J. L. Hodges, “Discriminatory analysis. Non-parametric discrimination: Consistency properties,” International Statistical Review/Revue Internationale de Statistique, 1989, pp. 238-247.
[20] R. Rusu and S. Cousins, “3D Is Here: Point Cloud Library (PCL),” In IEEE International Conference on Robotics and Automation (ICRA), 2011, pp. 1-4.
[21] R. W. Floyd, “Algorithm 97: Shortest path”, Communications of the ACM, Vol. 5, No. 6, Jun. 1962, p. 2. doi:10.1145/367766.368168

  
comments powered by Disqus

Copyright © 2018 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.