Abstract
The parallax artifacts introduced due to movement of objects across different views in the overlapping area drastically degrade the video stitching quality. To alleviate such visual artifacts, this paper extend our earlier video stitching framework [1] by replacing a deep learning based object detection algorithm for parallax detection, and an optical flow estimation algorithm for parallax correction. Given a set of multi-view overlapping videos, the geometric look-up tables (G-LUT) are generated by stitching a reference frame from the multi-view input videos, which map the input video frames to the panorama domain. We propose to use a deep learning based approach to detect the moving objects in the overlapping area to identify the G-LUT control points which get affected by parallax. To compute the optimal locations of these parallax affected G-LUT control points we propose to use patch-match based optical flow (CPM-flow). The adjustment of G-LUT control points in the overlapping area may cause some unwanted geometric distortions in the non-overlapping area. Therefore, the G-LUT control points in close proximity of moving objects are also updated to ensure the smooth geometric transition between the overlapping and the non-overlapping area. Experimental results on challenging video sequences with very narrow overlapping areas (\(\mathord \sim \)3% to \(\mathord \sim \)10%) demonstrate that video stitching framework with the proposed parallax minimization scheme can significantly suppress the parallax artifacts occurring due to the moving objects. In comparison to our previous work, the computational time is reduce by \(\mathord \sim \)26% with the proposed scheme, while the stitching quality is also marginally improved.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Kakli, M.U., Cho, Y., Seo, J.: Minimization of parallax artifacts in video stitching for moving foregrounds. IEEE Access 6, 57763â57777 (2018)
Szeliski, R.: Image alignment and stitching: a tutorial. Found. Trends Comput. Graph. Vis. 2(1), 1â104 (2006)
Lin, K., Jiang, N., Cheong, L.-F., Do, M., Lu, J.: SEAGULL: seam-guided local alignment for parallax-tolerant image stitching. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016, Part III. LNCS, vol. 9907, pp. 370â385. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46487-9_23
Gao, J, Kim, S. J., Brown, M. S.: Constructing image panoramas using dual-homography warping. In: Proceedings of the IEEE CVPR, pp. 49â56 (2011)
Zaragoza, J., Chin, T., Tran, Q., Brown, M.S., Suter, D.: As-projective-as-possible image stitching with moving DLT. IEEE TPAMI. 36(7), 1285â1298 (2014)
Chang, C.-H., Sato, Y., Chuang, Y.-Y.: Shape-preserving half projective warps for image stitching. In: IEEE CVPR, pp. 3254â3261 (2014)
Shimizu, T., Yoneyama, A., Takishima, Y.: A fast video stitching method for motion-compensated frames in compressed video streams. In: ICCE, pp. 173â174 (2006)
Zhang, F., Liu, F.: Parallax-tolerant image stitching. In: IEEE CVPR, pp. 3262â3269 (2014)
He, B., Yu, S.: Parallax-robust surveillance video stitching. Sensors 16(1), 7 (2015)
Jiang, W., Gu, J.: Video stitching with spatial-temporal content preserving warping. In: IEEE CVPRW, pp. 42â48 (2015)
He, B., Zhao, G., Liu, Q.: Panoramic video stitching in multi-camera surveillance system. In: IVCNZ, pp. 1â6 (2010)
Lin, K., Liu, S., Cheong, L.-F., Zeng, B.: Seamless video stitching from hand-held camera inputs. Comput. Graph. Forum 35(5), 479â487 (2016)
Guo, H., Liu, S., He, T., Zhu, S., Zeng, B., Gabbouj, M.: Joint video stitching and stabilization from moving cameras. IEEE TIP 25(11), 5491â5503 (2016)
Su, T., Nie, Y., Zhang, Z., Sun, H., Li, G.: Video stitching for handheld inputs via combined video stabilization. In: SIGGRAPH Asia Technical Briefs, pp. 25:1â25:4 (2016)
Nie, Y., Su, T., Zhang, Z., Sun, H., Li, G.: Dynamic video stitching via shakiness removing. IEEE TIP 27(1), 164â178 (2018)
Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE TIP 13(4), 600â612 (2004)
Qureshi, H.S., Khan, M.M., Hafiz, R., Cho, Y., Cha, J.: Quantitative quality assessment of stitched panoramic images. IET Image Process. 6(9), 1348â1358 (2012)
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: IEEE CVPR, pp. 779â788 (2016)
Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part V. LNCS, vol. 8693, pp. 740â755. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10602-1_48
Hu, Y., Song, R., Li, Y.: Efficient coarse-to-fine patchmatch for large displacement optical flow. In: IEEE CVPR, pp. 5704â5712 (2016)
Horn, B.K.P., Schunck, B.G.: Determining optical flow. Artificial Intelligence (1981)
Lucas, B., Kanade, T.: An iterative image registration technique with an application to stereo vision. In: IJCAI, pp. 674â679 (1981)
Video results of proposed scheme. http://sites.google.com/site/parallaxminconf2019. Accessed 29 Sept 2019
Acknowledgments
This work was supported by Korea Government (MSIT) 19ZR1120 (Development of 3D Spatial Media Core Technology).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
Âİ 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Kakli, M.U., Cho, Y., Seo, J. (2020). Parallax-Tolerant Video Stitching with Moving Foregrounds. In: Palaiahnakote, S., Sanniti di Baja, G., Wang, L., Yan, W. (eds) Pattern Recognition. ACPR 2019. Lecture Notes in Computer Science(), vol 12047. Springer, Cham. https://doi.org/10.1007/978-3-030-41299-9_49
Download citation
DOI: https://doi.org/10.1007/978-3-030-41299-9_49
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-41298-2
Online ISBN: 978-3-030-41299-9
eBook Packages: Computer ScienceComputer Science (R0)