Saliency-weighted graphs for efficient visual content description and their applications in real-time image retrieval systems
View/ Open
Mehmood_Journal_of_Real-Time_Image_Processing (1.930Mb)
Download
Publication date
2017-09Keyword
Attributed relational graphImage representation
Content-based image retrieval
Saliency map
Real-time retrieval
Rights
© Springer-Verlag Berlin Heidelberg 2015. Reproduced in accordance with the publisher's self-archiving policy. The final publication is available at Springer via https://doi.org/10.1007/s11554-015-0536-0.Peer-Reviewed
YesOpen Access status
openAccessAccepted for publication
2015-09-29
Metadata
Show full item recordAbstract
The exponential growth in the volume of digital image databases is making it increasingly difficult to retrieve relevant information from them. Efficient retrieval systems require distinctive features extracted from visually rich contents, represented semantically in a human perception-oriented manner. This paper presents an efficient framework to model image contents as an undirected attributed relational graph, exploiting color, texture, layout, and saliency information. The proposed method encodes salient features into this rich representative model without requiring any segmentation or clustering procedures, reducing the computational complexity. In addition, an efficient graph-matching procedure implemented on specialized hardware makes it more suitable for real-time retrieval applications. The proposed framework has been tested on three publicly available datasets, and the results prove its superiority in terms of both effectiveness and efficiency in comparison with other state-of-the-art schemes.Version
Accepted manuscriptCitation
Ahmad J, Sajjad M, Mehmood I et al (2017) Saliency-weighted graphs for efficient visual content description and their applications in real-time image retrieval systems. Journal of Real-Time Image Processing. 13(3): 431-447.Link to Version of Record
https://doi.org/10.1007/s11554-015-0536-0Type
Articleae974a485f413a2113503eed53cd6c53
https://doi.org/10.1007/s11554-015-0536-0