Research on Real time Acquisition System Based on Binocular Stereoscopic Vision
DOI:
https://doi.org/10.63313/AERpc.9063Keywords:
Real-time System, Binocular Stereoscopic Vision, 3D Acquisition, Stereo Matching, Depth MapAbstract
This paper presents the design and implementation of a real-time 3D acquisition system utilizing binocular stereoscopic vision. The primary objective is to develop a robust and efficient pipeline capable of capturing, processing, and reconstructing three-dimensional geometry of dynamic scenes with low latency. The core methodology integrates synchronized image capture from a calibrated stereo camera pair, followed by real-time stereo rectification and a dense stereo matching algorithm optimized for speed. The system successfully achieves real-time performance, generating dense depth maps at a frame rate sufficient for interactive applications. Experimental results demonstrate the system's accuracy in reconstructing static objects and its capability to track depth variations in moderately dynamic environments. The conclusion highlights that the implemented system provides a practical and effective solution for real-time 3D perception, establishing a reliable foundation for applications in robotics guidance, quality inspection, and augmented reality where immediate spatial feedback is critical.
References
[1] Zhang, Z. (2000). A flexible new technique for camera calibration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(11), 1330-1334. https://doi.org/10.1109/34.888718
[2] Hartley, R. I. (1997). In defense of the eight-point algorithm. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19(6), 580-593. https://doi.org/10.1109/34.601246
[3] Scharstein, D., & Szeliski, R. (2002). A taxonomy and evaluation of dense two-frame stereo correspondence algorithms. International Journal of Computer Vision, 47(1/3), 7-42. https://doi.org/10.1023/A:1014573219977
[4] Hirschmüller, H. (2008). Stereo processing by semiglobal matching and mutual infor-mation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 30(2), 328-341. https://doi.org/10.1109/TPAMI.2007.1166
[5] Kendall, A., Martirosyan, H., Dasgupta, S., Henry, P., Kennedy, R., Bachrach, A., & Bry, A. (2017). End-to-end learning of geometry and context for deep stereo regression. Proceed-ings of the IEEE International Conference on Computer Vision (ICCV), 66-75. https://doi.org/10.1109/ICCV.2017.17
[6] Guo, X., Yang, K., Yang, W., Wang, X., & Li, H. (2019). Group-wise correlation stereo net-work. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recogni-tion (CVPR), 3273-3282. https://doi.org/10.1109/CVPR.2019.00339
[7] Chang, J.-R., & Chen, Y.-S. (2018). Pyramid stereo matching network. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 5410-5418. https://doi.org/10.1109/CVPR.2018.00567
[8] Faugeras, O., & Lustman, F. (1988). Motion and structure from motion in a piecewise pla-nar environment. International Journal of Pattern Recognition and Artificial Intelligence, 2(3), 485-508. https://doi.org/10.1142/S0218001488000285
[9] Brown, M. Z., Burschka, D., & Hager, G. D. (2003). Advances in computational stereo. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(8), 993-1008. https://doi.org/10.1109/TPAMI.2003.1217603
[10] Mayer, N., Ilg, E., Häusser, P., Fischer, P., Cremers, D., Dosovitskiy, A., & Brox, T. (2016). A large dataset to train convolutional networks for disparity, optical flow, and scene flow es-timation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recogni-tion (CVPR), 4040-4048. https://doi.org/10.1109/CVPR.2016.438
[11] Menze, M., & Geiger, A. (2015). Object scene flow for autonomous vehicles. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 3061-3070. https://doi.org/10.1109/CVPR.2015.7298925
[12] Lipson, L., Teed, Z., & Deng, J. (2021). RAFT-Stereo: Multilevel recurrent field transforms for stereo matching. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2070-2079. https://doi.org/10.1109/ICCV48922.2021.00208
Downloads
Published
Issue
Section
License
Copyright (c) 2025 by author(s) and Erytis Publishing Limited.

This work is licensed under a Creative Commons Attribution 4.0 International License.








