Automatic Subgrouping of Multitrack Audio
Abstract
Subgrouping is a mixing technique where the outputs of a subset of audio tracks in a multitrack are summed to a single audio bus. This is done so that the mix engineer can apply signal processing to an entire subgroup, speed up the mix work flow and manipulate a number of audio tracks at once. In this work, we investigate which audio features from a set of 159 can be used to automatically subgroup multitrack audio. We determine a subset of audio features from the original 159 audio features to use for automatic subgrouping, by performing feature selection using a Random Forest classifier on a dataset of 54 individual multitracks. We show that by using agglomerative clustering on 5 test multitracks, the entire set of audio features incorrectly clusters 35.08% of the audio tracks, while the subset of audio features incorrectly clusters only 7.89% of the audio tracks. Furthermore, we also show that using the entire set of audio features, ten incorrect subgroups are created. However, when using the subset of audio features, only five incorrect subgroups are created. This indicates that our reduced set of audio features provides a significant increase in classification accuracy for the creation of subgroups automatically.
References (30)
- REFERENCES
- David Ronan, Brecht De Man, Hatice Gunes, and Joshua D. Reiss, "The impact of subgrouping practices on the percep- tion of multitrack mixes," in 139th Convention of the Audio Engineering Society, October 2015.
- Jeffrey Scott and Youngmoo E Kim, "Instrument identifi- cation informed multi-track mixing.," in 14th International Society for Music Information Retrieval Conference (ISMIR 2013), 2013, pp. 305-310.
- Jeffrey Scott, Matthew Prockup, Erik M Schmidt, and Youngmoo E Kim, "Automatic multi-track mixing using lin- ear dynamical systems," in Proceedings of the 8th Sound and Music Computing Conference, Padova, Italy, 2011.
- Joshua D Reiss, "Intelligent systems for mixing multichan- nel audio," in 17th International Conference on Digital Sig- nal Processing (DSP), 2011. IEEE, 2011, pp. 1-6.
- Enrique Perez-Gonzalez and Joshua D Reiss, "Automatic mixing," DAFX: Digital Audio Effects, Second Edition, pp. 523-549, 2011.
- Philippe Hamel, Sean Wood, and Douglas Eck, "Automatic identification of instrument classes in polyphonic and poly- instrument audio.," in 10th International Society for Music Information Retrieval Conference (ISMIR 2009), 2009, pp. 399-404.
- Judith C Brown, Olivier Houix, and Stephen McAdams, "Feature dependence in the automatic identification of mu- sical woodwind instruments," The Journal of the Acoustical Society of America, vol. 109, no. 3, pp. 1064-1072, 2001.
- Antti Eronen and Anssi Klapuri, "Musical instrument recog- nition using cepstral coefficients and temporal features," in Proceedings IEEE International Conference on Acoustics, Speech, and Signal Processing, 2000. ICASSP'00. IEEE, 2000, vol. 2, pp. II753-II756.
- Keith D Martin and Youngmoo E Kim, "Musical instrument identification: A pattern-recognition approach," The Jour- nal of the Acoustical Society of America, vol. 104, no. 3, pp. 1768-1768, 1998.
- Brecht De Man, Brett Leonard, Richard King, and Joshua D. Reiss, "An analysis and evaluation of audio features for mul- titrack music mixtures," in 15th International Society for Music Information Retrieval Conference (ISMIR 2014), Oc- tober 2014.
- Philippe Hamel, Simon Lemieux, Yoshua Bengio, and Dou- glas Eck, "Temporal pooling and multiscale learning for au- tomatic annotation and ranking of music audio.," in 12th International Society for Music Information Retrieval Con- ference (ISMIR 2011), 2011, pp. 729-734.
- Christian Uhle, "Tempo induction by investigating the met- rical structure of music using a periodicity signal that relates to the tatum period," in 6th International Society for Music Information Retrieval Conference (ISMIR 2005), 2005.
- Theodoros Giannakopoulos and Aggelos Pikrakis, Introduc- tion to Audio Analysis: A MATLAB R Approach, Academic Press, 2014.
- George Tzanetakis and Perry Cook, "Musical genre classi- fication of audio signals," IEEE transactions on Speech and Audio Processing, vol. 10, no. 5, pp. 293-302, 2002.
- Olivier Lartillot and Petri Toiviainen, "A matlab toolbox for musical feature extraction from audio," in International Con- ference on Digital Audio Effects, 2007, pp. 237-244.
- Geoffroy Peeters, "A large set of audio features for sound description (similarity and classification) in the cuidado project," 2004.
- Leo Breiman, "Random forests," Machine learning, vol. 45, no. 1, pp. 5-32, 2001.
- Carolin Strobl, Anne-Laure Boulesteix, Achim Zeileis, and Torsten Hothorn, "Bias in random forest variable importance measures: Illustrations, sources and a solution," BMC bioin- formatics, vol. 8, no. 1, pp. 25, 2007.
- Carolin Strobl, Anne-Laure Boulesteix, Thomas Kneib, Thomas Augustin, and Achim Zeileis, "Conditional variable importance for random forests," BMC bioinformatics, vol. 9, no. 1, pp. 307, 2008.
- Robin Genuer, Jean-Michel Poggi, and Christine Tuleau- Malot, "Variable selection using random forests," Pattern Recognition Letters, vol. 31, no. 14, pp. 2225-2236, 2010.
- Pavel Berkhin, "A survey of clustering data mining tech- niques," in Grouping multidimensional data, pp. 25-71. Springer, 2006.
- Stephen P Borgatti, "How to explain hierarchical clustering," 1994.
- Zhouyu Fu, Guojun Lu, Kai Ming Ting, and Dengsheng Zhang, "A survey of audio-based music classification and annotation," IEEE Transactions on Multimedia, vol. 13, no. 2, pp. 303-319, 2011.
- Tim Pohle, Elias Pampalk, and Gerhard Widmer, "Evalu- ation of frequently used audio features for classification of music into perceptual categories," in Proceedings of the Fourth International Workshop on Content-Based Multime- dia Indexing. Citeseer, 2005, vol. 162.
- Todor Ganchev, Nikos Fakotakis, and George Kokkinakis, "Comparative evaluation of various mfcc implementations on the speaker verification task," in Proceedings of the SPECOM, 2005, vol. 1, pp. 191-194.
- Michael Terrell, Joshua D Reiss, and Mark Sandler, "Au- tomatic noise gate settings for drum recordings containing bleed from secondary sources," EURASIP Journal on Ad- vances in Signal Processing, vol. 2010, pp. 10, 2010.
- Enrique Perez-Gonzalez and Joshua Reiss, "Automatic equalization of multichannel audio using cross-adaptive methods," in 127th Convention of the Audio Engineering Society. Audio Engineering Society, 2009.
- Michael Terrell, Andrew Simpson, and Mark Sandler, "The mathematics of mixing," Journal of the Audio Engineering Society, vol. 62, no. 1/2, pp. 4-13, 2014.
- Zheng Ma, Brecht De Man, Pedro Duarte Pestana, Dawn A. A. Black, and Joshua D. Reiss, "Intelligent multitrack dynamic range compression," Journal of the Audio Engi- neering Society, 2015.