Key research themes
1. How are digital audio effects designed and modeled to manipulate and enhance musical sound?
This research area investigates the design, modeling, and implementation of digital audio effects (DAFx) that manipulate musical sound for creative and production purposes. It focuses on algorithmic development, real-time processing, and the application of machine learning for black-box and parametric modeling of audio effects such as equalization, amplification, and non-linear distortions. These advances are significant for music production, mixing, live performance, and the emulation of analog hardware in digital environments.
2. What advances enable automatic detection and removal of non-speech vocal sounds (e.g., breath sounds) in digital audio recordings?
This important subfield addresses the challenge of identifying and eliminating unwanted non-speech vocal sounds such as breaths that may detract from the clarity of vocal recordings in music, broadcasting, and speech processing. The focus encompasses algorithmic approaches employing signal processing and deep learning methods to automate this laborious process. Effective solutions improve both production efficiency and audio quality, with notable progress shown in attention U-Net architectures that combine accuracy with reduced computational complexity.
3. How are digital sound synthesis and digital sound reconstruction advancing in generating realistic audio?
This theme covers algorithmic innovations in synthesizing sounds digitally, particularly through physical modeling and digital reconstruction techniques. The goal is to produce realistic and high-fidelity synthetic musical or acoustic sounds. Physical modeling simulates vibrating physical structures, enabling expressive synthesis linked to instrument mechanics. Digital sound reconstruction (DSR) techniques aim to overcome limitations of traditional loudspeaker designs, especially in low-frequency domains, using array configurations and shutter gate mechanisms to enhance sound pressure output. These approaches enable real-time synthesis and new sound generation paradigms critical for musical instrument digital interfaces and audio reproduction devices.