Material Detail

Hierarchical Bayesian Models for Audio and Music Processing

Hierarchical Bayesian Models for Audio and Music Processing

This video was recorded at NIPS Workshop on Music, Brain and Cognition, Whistler 2007 . In recent years, there has been an increasing interest in statistical approaches and tools from machine learning for the analysis of audio and music signals, driven partially by applications in music information retrieval, computer aided music education and interactive music performance systems. The application of statistical techniques is quite natural: acoustical time series can be conveniently modelled using hierarchical signal models by incorporating prior knowledge from various sources: from physics or studies of human cognition and perception. Once a realistic hierarchical model is constructed, many audio processing tasks such as coding, restoration, transcription, separation, identification or resynthesis can be formulated consistently as Bayesian posterior inference problems. In this talk, we will review recent advances in various signal models for audio and music signal analysis. In particular, factorial switching state space models, Gamma-Markov random fields will be discussed. Some models admit exact inference, otherwise efficient algorithms based on variational or stochastic approximation methods can be developed. We will illustrate applications on music transcription, tempo tracking, restoration and source separation applications.


  • User Rating
  • Comments
  • Learning Exercises
  • Bookmark Collections
  • Course ePortfolios
  • Accessibility Info

More about this material


Log in to participate in the discussions or sign up if you are not already a MERLOT member.