Please use this identifier to cite or link to this item:
Title: Automatic annotation of image databases based on implicit crowdsourcing, visual concept modeling and evolution
Authors: Tsapatsoulis, Nicolas 
Ntalianis, Klimis S. 
Doulamis, Anastasios D. 
Matsatsinis, Nikolaos F.
Keywords: Implicit crowdsourcing;User feedback;Visual concept modeling;Clickthrough data;Automatic image annotation
Category: Media and Communications
Field: Social Sciences
Issue Date: Mar-2014
Publisher: Springer US
Source: Multimedia Tools and Applications, 2014, Volume 69, Issue 2, pages 397-421
Abstract: In this paper a novel approach for automatically annotating image databases is proposed. Despite most current schemes that are just based on spatial content analysis, the proposed method properly combines several innovative modules for semantically annotating images. In particular it includes: (a) a GWAP-oriented interface for optimized collection of implicit crowdsourcing data, (b) a new unsupervised visual concept modeling algorithm for content description and (c) a hierarchical visual content display method for easy data navigation, based on graph partitioning. The proposed scheme can be easily adopted by any multimedia search engine, providing an intelligent way to even annotate completely non-annotated content or correct wrongly annotated images. The proposed approach currently provides very interesting results in limited-size both standard and generic datasets and it is expected to add significant value especially to billions of non-annotated images existing in the Web. Furthermore expert annotators can gain important knowledge relevant to user new trends, language idioms and styles of searching.
ISSN: 1380-7501
DOI: 10.1007/s11042-012-0995-2
Rights: © Springer Science+Business Media, LLC 2012
Type: Article
Appears in Collections:Άρθρα/Articles

Show full item record

Citations 10

checked on Jun 23, 2019


checked on Jul 20, 2019

Page view(s)

Last Week
Last month
checked on Jul 23, 2019

Google ScholarTM



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.