Variable block size motion estimation implementation on compute unified device architecture (CUDA) | IEEE Conference Publication | IEEE Xplore

Variable block size motion estimation implementation on compute unified device architecture (CUDA)


Abstract:

This paper proposes a highly parallel variable block size full search motion estimation algorithm with concurrent parallel reduction (CPR) on graphics processing unit (GP...Show More

Abstract:

This paper proposes a highly parallel variable block size full search motion estimation algorithm with concurrent parallel reduction (CPR) on graphics processing unit (GPU) using compute unified device architecture (CUDA). This approach minimizes memory access latency by using high-speed on-chip memory of GPU. By applying parallel reductions concurrently depending on the amount of data and the data dependency, the proposed approach increases thread utilization and decreases the number of synchronization points which cause latency. Experimental results show that the proposed approach achieves substantial improvement up to 92 times than the central processing unit (CPU) only counterpart.
Date of Conference: 11-14 January 2013
Date Added to IEEE Xplore: 28 March 2013
ISBN Information:

ISSN Information:

Conference Location: Las Vegas, NV, USA

References

References is not available for this document.