Large-scale duplicate image retrieval technical research for the internet

For the typical social media application on the internet,a large-scale distributed duplicate image retrieval ap-proach based on random projection and the block DCT coefficients was proposed.On the basis of Hadoop,this approach exploited image signatures generated by random projection mapping to retr...

Full description

Saved in:
Bibliographic Details
Main Authors: Shu-peng WANG, Ming CHEN, Guang-jun WU
Format: Article
Language:zho
Published: Editorial Department of Journal on Communications 2014-12-01
Series:Tongxin xuebao
Subjects:
Online Access:http://www.joconline.com.cn/zh/article/doi/10.3969/j.issn.1000-436x.2014.12.023/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:For the typical social media application on the internet,a large-scale distributed duplicate image retrieval ap-proach based on random projection and the block DCT coefficients was proposed.On the basis of Hadoop,this approach exploited image signatures generated by random projection mapping to retrieve HBase efficiently.And candidate images with high-recall were achieved.Then in order to improve the retrieval precision,the block DCT coefficients were used to further filter candidate images.For 12 million images,experimental results showed that with our approach the recall ratio reached 98%,the precision ratio reached 93.2%,and the average retrieval time was 6.7s when H=2 and T=150.
ISSN:1000-436X