Skip to main content
Log in

Minimal vision system for linear camera self-calibration and 3D reconstruction using translation platform

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Camera calibration and 3D reconstruction are two crucial steps in computer vision. With the progress of robot and autonomous rover systems, zooming cameras are widely applied, where online calibration and 3D reconstruction are becoming more and more important. This paper proposed a minimal vision system that consists of a translation platform and an un-calibrated camera mounted on. With this minimal system, we can linearly calibrate the camera online and reconstruct 3D structures from three un-calibrated images by utilizing translation motions. The two images generated by translating the camera to allow the recovery of scene depths. Depths are then analyzed by error analysis models and are utilized to determine the infinite homography between the third image and any of the two translated images. The intrinsic parameters are then calibrated linearly from the computed infinite homography. Camera motion estimation and 3D reconstruction are then readily determined from the intrinsic calibration results. We also proposed a two-step optimization method to refine both the calibration and 3D reconstruction results by minimizing the overall back-projection errors across the three images within a tiny-scale bundle adjustment framework. The proposed method has been validated with both simulation and real image data. The results demonstrate that the proposed minimal linear system can solve the online camera self-calibration problem and the reconstruction of 3D structure. The paper suggests a framework of minimal linear system by using an un-calibrated camera and a low-cost translation platform to address the problems of linear self-calibration of camera, motion estimation, 3D reconstruction, and optimization, in a practical, easy, and accurate way.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

References

  1. Bay H, Ess A, Tuytelaars T, Gool LV (2008) SURF: speeded up robust features. Comput Vis Image Underst 110(3):346–359

    Article  Google Scholar 

  2. Filippo B, Menegatti E, Pretto A (2018) Robust intrinsic and extrinsic calibration of RGB-D cameras. IEEE Trans Robot 34(5):1315–1332

    Article  Google Scholar 

  3. Fischler MA, Bolles RC (1981) Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun ACM 24(6):381–395

    Article  MathSciNet  Google Scholar 

  4. Ge K, Hu H, Feng J, Zhou Z (2016) Depth estimation using a sliding camera. IEEE Trans Image Process 25(2):726–739

    Article  MathSciNet  MATH  Google Scholar 

  5. Hartley RI (1994) Self-calibration from multiple views with a rotating camera. The Third European Conference on Computer Vision. Springer, Berlin, Heidelberg

  6. Hartley R, Zisserman A (2004) Multiple view geometry in computer vision, 2nd edn. Cambridge University Press, Cambridge

    Book  MATH  Google Scholar 

  7. Houssineau J, Clark DE, Ivekovic S, Lee CS, Franco J (2016) A unified approach for multi-object triangulation, tracking and camera calibration. IEEE Trans Signal Process 64(11):2934–2948

    Article  MathSciNet  MATH  Google Scholar 

  8. Hu Z, Matsuyama T (2015) Bayesian perspective-plane (BPP) with maximum likelihood searching for visual localization. Multimed Tools Appl 74(21):9547–9572

    Article  Google Scholar 

  9. Hu Z, Tan Z (2007) Depth recovery and affine reconstruction under camera pure translation. Pattern Recogn 40(10):2826–2836

    Article  MATH  Google Scholar 

  10. Hui B, Wen G, Zhang P, Li D (2013) A novel line scan camera calibration technique with an auxiliary frame camera. IEEE Trans Instrum Meas 62(9):2567–2575

    Article  Google Scholar 

  11. Kottari K, Delibasis K, Plagianakos V (2017) Real time vision-based measurements for quality control of industrial rods on a moving conveyor. Multimed Tools Appl 77(6):1–18

    Google Scholar 

  12. Liu J, Li Y, Chen S (2014) Robust camera calibration by optimal localization of spatial control points. IEEE Trans Instrum Meas 63(12):3076–3087

    Article  Google Scholar 

  13. Liu Y, Nie L, Han L, Zhang L, Rosenblum DS (2015) Action2Activity: recognizing complex activities from sensor data. In: International joint conferences on artificial intelligence, pp 1617–1623

  14. Liu Y, Nie LQ, Liu L, Rosenblum DS (2016) From action to activity: sensor-based activity recognition. Neurocomputing 181(99):108–115

    Article  Google Scholar 

  15. Liu Y, Zhang L, Nie L, Yan Y, Rosenblum DS (2016) Fortune teller: predicting your career path. In: Association for the advancement of artificial intelligence, pp 201–207

  16. Lourakis MIA, Argyros AA (2009) SBA: a software package for generic sparse bundle adjustment. ACM Trans Math Softw 36(1):1–30

    Article  MathSciNet  MATH  Google Scholar 

  17. Ma SD (1996) A self-calibration technique for active vision systems. IEEE Trans Robot Autom 12(1):114–120

    Article  MathSciNet  Google Scholar 

  18. Maybank SJ, Faugeras OD (1992) A theory of self-calibration of a moving camera. Int J Comput Vis 8(2):123–151

    Article  Google Scholar 

  19. Panahandeh G, Jansson M, Handel P (2015) Calibration of an IMU-camera cluster using planar mirror reflection and its observability analysis. IEEE Trans Instrum Meas 64(1):75–88

    Article  Google Scholar 

  20. Ramalingam S, Sturm P (2016) A unifying theory for camera calibration. IEEE Trans Pattern Anal Mach Intell 39(7):1309–1319

    Article  Google Scholar 

  21. Vasconcelos F, Barreto JP, Boyer E (2017) Automatic camera calibration using multiple sets of pair-wise correspondences. IEEE Trans Pattern Anal Mach Intell PP(99):1

    Google Scholar 

  22. Wang L, Kang SB, Shum HY, Xu GY (2004) Error analysis of pure rotation-based self-calibration. IEEE Trans Pattern Anal Mach Intell 26:275–280

    Article  Google Scholar 

  23. Wang Q, Liu Y, Shen Y (2011) An accurate extrinsic camera self-calibration method in non-overlapping camera sensor networks. In: IEEE conf. on instrumentation and measurement technology conference (I2MTC), pp 1–6

  24. Xue JP, Su XY, Xiang LQ, Chen WJ (2012) Using concentric circles and wedge grating for camera calibration. Appl Opt 51(17):3811–3816

    Article  Google Scholar 

  25. Zhang Z (2000) A flexible new technique for camera calibration. IEEE Trans Pattern Anal Mach Intell 22(11):1330–1334

    Article  Google Scholar 

  26. Zhao B, Hu Z (2015) Camera self-calibration from translation by referring to a known camera. Appl Opt 54(30):7789–7798

    Article  Google Scholar 

  27. Zhao Y, Li X, Li W (2012) Binocular vision system calibration based on a one-dimensional target. Appl Opt 51(16):3338–3345

    Article  Google Scholar 

Download references

Acknowledgements

The work presented in this paper was funded by the National Natural Science Foundation of China (NSFC) (No. 51679181), National Key R&D Program of China (2018YFB1600801), the Major Project of Technical Innovation of Hubei Province (No. 2016AAA007), and the Science-technology Programs Prior Funds for Overseas Chinese Talents of Hubei Province (2016-12).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhaozheng Hu.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hu, Z., Zhao, B. Minimal vision system for linear camera self-calibration and 3D reconstruction using translation platform. Multimed Tools Appl 78, 25731–25751 (2019). https://doi.org/10.1007/s11042-019-7666-5

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-019-7666-5

Keywords

Navigation