Multimodal music retrieval for large databases

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

5 Scopus citations

Abstract

In this contribution we present a novel multi-modal access to large MP3 music data-bases. Retrieval can be either fulfilled in a content-based manner or by keywords. As input modalities speech by natural language utterances or singing, and manual interaction by handwriting, typing or hardkeys are used. In order to achieve especially robust retrieval results and automatically suggest music to the user contextual knowledge of the time, date, season, user emotion, and listening habits is integrated in the retrieval process. The system communicates with the user by speech or visual reactions. The concepts shown are especially designed for home and mobile access on Tablet-PCs, PDAs, and similar PC solutions. The paper discusses the concept and a working prototype called Shangrila. An evaluation by a user study leads to an impression of the capabilities of the suggested approach to multimodal music retrieval.

Original languageEnglish
Title of host publication2004 IEEE International Conference on Multimedia and Expo (ICME)
Pages755-758
Number of pages4
StatePublished - 2004
Event2004 IEEE International Conference on Multimedia and Expo (ICME) - Taipei, Taiwan, Province of China
Duration: 27 Jun 200430 Jun 2004

Publication series

Name2004 IEEE International Conference on Multimedia and Expo (ICME)
Volume2

Conference

Conference2004 IEEE International Conference on Multimedia and Expo (ICME)
Country/TerritoryTaiwan, Province of China
CityTaipei
Period27/06/0430/06/04

Fingerprint

Dive into the research topics of 'Multimodal music retrieval for large databases'. Together they form a unique fingerprint.

Cite this