Authors:
Tran Phuong Nhung
1
;
Cam-Tu Nguyen
2
;
Jinhee Chun
1
;
Ha Vu Le
3
and
Takeshi Tokuyama
1
Affiliations:
1
Tohoku University, Japan
;
2
Nanjing University, China
;
3
VNU University of Engineering and Technology, Vietnam
Keyword(s):
Visual Saliency, Image Annotation, Multiple Instance Learning.
Related
Ontology
Subjects/Areas/Topics:
Artificial Intelligence
;
Computational Intelligence
;
Data Reduction and Quality Assessment
;
Evolutionary Computing
;
Knowledge Discovery and Information Retrieval
;
Knowledge-Based Systems
;
Machine Learning
;
Mining High-Dimensional Data
;
Mining Multimedia Data
;
Soft Computing
;
Symbolic Systems
;
Visual Data Mining and Data Visualization
Abstract:
This paper presents a novel approach to image annotation based on multi-instance learning (MIL) and saliency map. Image Annotation is an automatic process of assigning labels to images so as to obtain semantic retrieval of images. This problem is often ambiguous as a label is given to the whole image while it may only corresponds to a small region in the image. As a result, MIL methods are suitable solutions to resolve the ambiguities during learning. On the other hand, saliency detection aims at detecting foreground/background regions in images. Once we obtain this information, labels and image regions can be aligned better, i.e., foreground labels (background labels) are more sensitive to foreground areas (background areas). Our proposed method, which is based on an ensemble of MIL classifiers from two views (background/foreground), improves annotation performance in comparison to baseline methods that do not exploit saliency information.