Proceedings of 3rd International Conference on Multimedia Technology(ICMT-13)

Recover 3D Information of the Moving Object from Video Streams

Authors
Zheng Yu-tong, Li Ming, Liao Fang
Corresponding Author
Zheng Yu-tong
Available Online November 2013.
DOI
https://doi.org/10.2991/icmt-13.2013.202How to use a DOI?
Keywords
Binocular, corresponding points, epipolar constraints, sparse depth map, autonomous navigation
Abstract
Perception of the moving object in 3D from video streams has been one hot topic in computer vision. We present a fast method to reconstruct 3D information of the moving object from binocular video streams. System is assembled as two pipelines, technica are used to excavate the potential parallelism. With the corresponding points searching confined to very limited and credible region, the mismatching errors and time-consumed computation are reduced considerably. At the last, sparse depth map is calculated and then 3D contours and location of the object are estimated. The system is implemented and tested with outdoor and indoor moving object perception on 640×480 frame. Results show that the proposed method is improved in speed and stability. It can be used as a reference for autonomous navigation of mobile robot and object tracking.
Open Access
This is an open access article distributed under the CC BY-NC license.

Download article (PDF)

Proceedings
3rd International Conference on Multimedia Technology(ICMT-13)
Part of series
Advances in Intelligent Systems Research
Publication Date
November 2013
ISBN
978-90-78677-89-5
ISSN
1951-6851
DOI
https://doi.org/10.2991/icmt-13.2013.202How to use a DOI?
Open Access
This is an open access article distributed under the CC BY-NC license.

Cite this article

TY  - CONF
AU  - Zheng Yu-tong
AU  - Li Ming
AU  - Liao Fang
PY  - 2013/11
DA  - 2013/11
TI  - Recover 3D Information of the Moving Object from Video Streams
BT  - 3rd International Conference on Multimedia Technology(ICMT-13)
PB  - Atlantis Press
SN  - 1951-6851
UR  - https://doi.org/10.2991/icmt-13.2013.202
DO  - https://doi.org/10.2991/icmt-13.2013.202
ID  - Yu-tong2013/11
ER  -