Background music is often used to generate a specific atmosphere or to draw our attention to specific events. For example in movies or computer games it is often the accompanying music that conveys the emotional state of a scene and plays an important role for immersing the viewer or player into the virtual environment. In view of home-made videos, slide shows, and other consumer-generated visual media streams, there is a need for computer-assisted tools that allow users to generate aesthetically appealing music tracks in an easy and intuitive way. In this contribution, we consider a data-driven scenario where the musical raw material is given in form of a database containing a variety of audio recordings. Then, for a given visual media stream, the task consists in identifying, manipulating, overlaying, concatenating, and blending suitable music clips to generate a music stream that satisfies certain constraints imposed by the visual data stream and by user specifications. It is our main goal to give an overview of various content-based music processing and retrieval techniques that become important in data-driven sound track generation. In particular, we sketch a general pipeline that highlights how the various techniques act together and come into play when generating musically plausible transitions between subsequent music clips.
@InCollection{muller_et_al:DFU.Vol3.11041.175, author = {M\"{u}ller, Meinard and Driedger, Jonathan}, title = {{Data-Driven Sound Track Generation}}, booktitle = {Multimodal Music Processing}, pages = {175--194}, series = {Dagstuhl Follow-Ups}, ISBN = {978-3-939897-37-8}, ISSN = {1868-8977}, year = {2012}, volume = {3}, editor = {M\"{u}ller, Meinard and Goto, Masataka and Schedl, Markus}, publisher = {Schloss Dagstuhl -- Leibniz-Zentrum f{\"u}r Informatik}, address = {Dagstuhl, Germany}, URL = {https://drops.dagstuhl.de/entities/document/10.4230/DFU.Vol3.11041.175}, URN = {urn:nbn:de:0030-drops-34725}, doi = {10.4230/DFU.Vol3.11041.175}, annote = {Keywords: Sound track, content-based retrieval, audio matching, time-scale modification, warping, tempo, beat tracking, harmony} }
Feedback for Dagstuhl Publishing