This position paper deals with queries beyond text, mixing several multimedia contents: audio, video, image and text. Search approaches combining some of these formats have been studied, including query by example techniques in situations where only one format is considered. It is worth mentioning that most of these research works do not deal with text content. A new approach to allow users introducing multimodal queries and exploring multimedia repositories is proposed. For this purpose, different ranked result lists must be combined to produce the final results shown for a given query. The main goal of this proposal is to reduce the semantic gap between low level features and high level concepts in multimedia contents. The use of qualitative data giving more relevance to text content along with machine learning methods to combine results of monomodal retrieval systems is proposed. Although it is too soon to show experimentation results, a prototype implementing the approach is under development and evaluation. © 2012 ICST Institute for Computer Science, Social Informatics and Telecommunications Engineering.
CITATION STYLE
Martínez, Á., Lana Serrano, S., Martínez-Fernández, J. L., & Martínez, P. (2012). Multimodal queries to access multimedia information sources: First steps. In Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering (Vol. 60 LNICST, pp. 35–40). https://doi.org/10.1007/978-3-642-35145-7_5
Mendeley helps you to discover research relevant for your work.