![](/images/graphics-bg.png)
Frame Interpolation Based on Visual Correspondence and Coherency Sensitive Hashing
Joint Authors
Zi, Lingling
Wang, Qian
Du, Junping
Source
Mathematical Problems in Engineering
Issue
Vol. 2013, Issue 2013 (31 Dec. 2013), pp.1-11, 11 p.
Publisher
Hindawi Publishing Corporation
Publication Date
2013-07-09
Country of Publication
Egypt
No. of Pages
11
Main Subjects
Abstract EN
The technology of frame interpolation can be applied in intelligent monitoring systems to improve the quality of surveillance video.
In this paper, a region-guided frame interpolation algorithm is proposed by introducing two innovative improvements.
On the one hand, a detection approach is presented based on visual correspondence for detecting the motion regions that correspond to attracted objects in video sequences, which can narrow the prediction range of interpolated frames.
On the other hand, spatial and temporal mapping rules are proposed using coherency sensitive hashing, which can obtain more accurate predicted values of interpolated pixels.
Experiments show that the proposed method can achieve encouraging performance in terms of visual quality and quantitative measures.
American Psychological Association (APA)
Zi, Lingling& Du, Junping& Wang, Qian. 2013. Frame Interpolation Based on Visual Correspondence and Coherency Sensitive Hashing. Mathematical Problems in Engineering،Vol. 2013, no. 2013, pp.1-11.
https://search.emarefa.net/detail/BIM-1031898
Modern Language Association (MLA)
Zi, Lingling…[et al.]. Frame Interpolation Based on Visual Correspondence and Coherency Sensitive Hashing. Mathematical Problems in Engineering No. 2013 (2013), pp.1-11.
https://search.emarefa.net/detail/BIM-1031898
American Medical Association (AMA)
Zi, Lingling& Du, Junping& Wang, Qian. Frame Interpolation Based on Visual Correspondence and Coherency Sensitive Hashing. Mathematical Problems in Engineering. 2013. Vol. 2013, no. 2013, pp.1-11.
https://search.emarefa.net/detail/BIM-1031898
Data Type
Journal Articles
Language
English
Notes
Includes bibliographical references
Record ID
BIM-1031898