Audio-visual feature aggregation for query generation

Ren, R., Halvey, M. and Jose, J. (2009) Audio-visual feature aggregation for query generation. In: IEEE International Conference on Multimedia and Expo, 2009. ICME 2009, New York, NY , U.S.A., 28 Jun - 3 Jul 2009,

Full text not currently available from Enlighten.

Abstract

Using multiple examples has become a popular query scenario in multimedia retrieval. This paper explores a unified representation which accumulates various features from different examples to denote a query. Continuous low-level features are quantised into a set of discrete variants. These variants follow a similar distribution as text terms do in a given document collection. Three criteria are compared to justify this projection, including minimised chi-square, maximised entropy and minimised AC/DC. Statistics similar to text term frequency are computed from these variants for document similarity ranking. Two ranking functions, KL divergence and BM25, are used for multimedia retrieval. The evaluation collection consists of the Corel image set and TRECVid 2006 collection with four low-level visual features. Experimental results show that the overall query performance based on this representation is comparable and in some cases out-performs direct visual feature comparison and the K-median clustering.

Item Type:Conference Proceedings
Status:Published
Refereed:Yes
Glasgow Author(s) Enlighten ID:Jose, Professor Joemon and Halvey, Dr Martin and Ren, Dr Reede
Authors: Ren, R., Halvey, M., and Jose, J.
Subjects:Q Science > QA Mathematics > QA75 Electronic computers. Computer science
College/School:College of Science and Engineering > School of Computing Science

University Staff: Request a correction | Enlighten Editors: Update this record