TY - GEN
T1 - Image annotations by combining multiple evidence & WordNet
AU - Jin, Yohan
AU - Khan, Latifur
AU - Wang, Lei
AU - Awad, Mamoun
PY - 2005
Y1 - 2005
N2 - The development of technology generates huge amounts of non-textual information, such as images. An efficient image annotation and retrieval system is highly desired. Clustering algorithms make it possible to represent visual features of images with finite symbols. Based on this, many statistical models, which analyze correspondence between visual features and words and discover hidden semantics, have been published. These models improve the annotation and retrieval of large image databases. However, current state of the art including our previous work produces too many irrelevant keywords for images during annotation. In this paper, we propose a novel approach that augments the classical model with generic knowledge-based, WordNet. Our novel approach strives to prune irrelevant keywords by the usage of WordNet. To identify irrelevant keywords, we investigate various semantic similarity measures between keywords and finally fuse outcomes of all these measures together to make a final decision using Dempster-Shafer evidence combination. We have implemented various models to link visual tokens with keywords based on knowledge-based, WordNet and evaluated performance using precision, and recall using benchmark dataset. The results show that by augmenting knowledge-based with classical model we can improve annotation accuracy by removing irrelevant keywords.
AB - The development of technology generates huge amounts of non-textual information, such as images. An efficient image annotation and retrieval system is highly desired. Clustering algorithms make it possible to represent visual features of images with finite symbols. Based on this, many statistical models, which analyze correspondence between visual features and words and discover hidden semantics, have been published. These models improve the annotation and retrieval of large image databases. However, current state of the art including our previous work produces too many irrelevant keywords for images during annotation. In this paper, we propose a novel approach that augments the classical model with generic knowledge-based, WordNet. Our novel approach strives to prune irrelevant keywords by the usage of WordNet. To identify irrelevant keywords, we investigate various semantic similarity measures between keywords and finally fuse outcomes of all these measures together to make a final decision using Dempster-Shafer evidence combination. We have implemented various models to link visual tokens with keywords based on knowledge-based, WordNet and evaluated performance using precision, and recall using benchmark dataset. The results show that by augmenting knowledge-based with classical model we can improve annotation accuracy by removing irrelevant keywords.
KW - Corel Dataset
KW - Dempster-Shafer Rule
KW - Image Annotation
KW - Management
KW - Semantic-Similarity
KW - WordNet
UR - http://www.scopus.com/inward/record.url?scp=84883103974&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84883103974&partnerID=8YFLogxK
U2 - 10.1145/1101149.1101305
DO - 10.1145/1101149.1101305
M3 - Conference contribution
AN - SCOPUS:84883103974
SN - 1595930442
SN - 9781595930446
T3 - Proceedings of the 13th ACM International Conference on Multimedia, MM 2005
SP - 706
EP - 715
BT - Proceedings of the 13th ACM International Conference on Multimedia, MM 2005
T2 - 13th ACM International Conference on Multimedia, MM 2005
Y2 - 6 November 2005 through 11 November 2005
ER -