Large Scale Tag Recommendation Using Different Image Representations
Published in In the proceedings of Semantic Multimedia, 2009
Recommended citation: Rabeeh Abbasi, Marcin Grzegorzek, Steffen Staab, "Large Scale Tag Recommendation Using Different Image Representations." In the proceedings of Semantic Multimedia, 2009. https://doi.org/10.1007/978-3-642-10543-2_8
Nowadays, geographical coordinates (geo-tags), social annotations (tags), and low-level features are available in large image datasets. In our paper, we exploit these three kinds of image descriptions to suggest possible annotations for new images uploaded to a social tagging system. In order to compare the benefits each of these description types brings to a tag recommender system on its own, we investigate them independently of each other. First, the existing data collection is clustered separately for the geographical coordinates, tags, and low-level features. Additionally, random clustering is performed in order to provide a baseline for experimental results. Once a new image has been uploaded to the system, it is assigned to one of the clusters using either its geographical or low-level representation. Finally, the most representative tags for the resulting cluster are suggested to the user for annotation of the new image. Large-scale experiments performed for more than 400,000 images compare the different image representation techniques in terms of precision and recall in tag recommendation.