Professional Documents
Culture Documents
Project Information
Deliverables
Other (9)
1 of 19
Final release of the software tool for automatically annotating music pieces using
semantic concepts from the Audio Commons Ontology. This version of the tool
contains the improvements included after the second evaluation cycle.
API specification
Specification of the end-user Web APIs and internal audio analysis and
processing APIs.
Evaluation report on the first prototype tool for the automatic semantic description of music samples
Report on analysis results of the accuracy and effectiveness of the first prototype
for the automatic annotation of music samples. Identification of key areas for
2 of 19
improvement/extension.
Evaluation report on the first prototype tool for the automatic semantic description of music pieces
Report on analysis results of the accuracy and effectiveness of the first prototype
for the automatic annotation of music piece. Identification of key areas for
improvement/extension.
Evaluation report on the second prototype tool for the automatic semantic description of music pieces
Report on analysis results of the accuracy and effectiveness of the second
prototype for the automatic annotation of music pieces. Identification of key areas
3 of 19
for timbral analysis improvement/extension.
Evaluation report on the second prototype tool for the automatic semantic description of music
samples
Report on analysis results of the accuracy and effectiveness of the second
prototype for the automatic annotation of music samples. Identification of key
areas for improvement/extension.
Dissemination plan
Dissemination plan covering the dissemination of AudioCommons RTD
outcomes. This deliverable is transversal to several tasks of this work package.
Report on the analysis and compilation of state-of-the-art methods for the automatic annotation of
music pieces and music samples
This deliverable will include a document listing existing methods for several audio
characterisation tasks (related to music pieces and music samples) along with
pointers into already existing implementations or specifically developed
implementations that can be used in AudioCommons.
Report on the evaluation of the ACE from a holistic and technological perspective
Report on the evaluation of our system assessing how our tools support creative
work in specific domains and how the system facilitates the reuse of audio
content as a whole.
First prototype of timbral characterisation tool for semantically annotating non-musical content
5 of 19
Release of the first prototype of the software tool for the automatic annotation of
non-musical audio content.
First prototype tool for the automatic semantic description of music pieces
Release of the first prototype of the software tool for the automatic annotation of
music pieces.
Second prototype tool for the automatic semantic description of music pieces
Release of the second prototype of the software tool for the automatic annotation
of music pieces. This version of the tools contains improvements included after
the first evaluation cycle.
First prototype tool for the automatic semantic description of music samples
Release of the first prototype of the software tool for the automatic annotation of
music samples.
Second prototype of timbral characterisation tool for semantically annotating non-musical content
Release of the second prototype of the software tool for the automatic annotation
of non-musical audio content. This version of the tools contains improvements
included after the first evaluation cycle.
Second prototype tool for the automatic semantic description of music samples
Release of the second prototype of the software tool for the automatic annotation
of music samples. This version of the tools contains improvements included after
the first evaluation cycle.
Project website
A multimedia website with public information on the project and its evolution,
together with a private space for partners only as a discussion space and
document repository.
Publications
6 of 19
Mood Visualiser: Augmented Music Visualisation Gauging Audience Arousal
Author(s): Subramaniam, A., Barthet, M.
Published in: Proc. of the 12th International Audio Mostly Conference on
Augmented and Participatory Sound and Music Experiences, 2017
Acoustic scene classification by ensembling gradient boosting machine and convolutional neural
networks
Author(s): Fonseca, Eduardo; Gong, Rong; Bogdanov, Dmitry; Slizovskaia,
Olga; Gómez Gutiérrez, Emilia, 1975-; Serra, Xavier
Published in: Proc. of the Detection and Classification of Acoustic Scenes and
Events Workshop (DCASE), Issue 12, 2017
The MediaEval 2017 AcousticBrainz Genre Task: Content-based Music Genre Recognition from
Multiple Sources
Author(s): Bogdanov, Dmitry; Porter, Alastair; Urbano Merino, J.; Schreiber,
Hendrik; Gravier, Guillaume; Bischke, Benjamin; Demarty, Claire-Hélène;
Zaharieva, Maia; Riegler, Michael; Dellandrea, Emmanuel; Bogdanov, Dmitry;
Sutcliffe, Richard; Jones, Gareth J.F.; Larson, Martha
Published in: MediaEval 2017. Dublin, Ireland, Issue 10, 2017
Quantifying music tends and facts using editorial metadata from the Discogs database
Author(s): Bogdanov, Dmitry; Serra, Xavier
Published in: Proc. of the International Society for Music Information Retrieval
Conference (ISMIR), Issue 9, 2017
A Comparison of Audio Signal Preprocessing Methods for Deep Neural Networks on Music Tagging
Author(s): Choi, Keunwoo; Fazekas, György; Cho, Kyunghyun; Sandler, Mark
Published in: Proc. of the 26th European Signal Processing Conference
(EUSIPCO), Issue 6, 2018
7 of 19
General-purpose Tagging of Freesound Audio with AudioSet Labels: Task Description, Dataset, and
Baseline
Author(s): Fonseca, Eduardo; Plakal, Manoj; Font, Frederic; Ellis, Daniel P. W.;
Favory, Xavier; Pons, Jordi; Serra, Xavier
Published in: Proc. of the Detection and Classification of Acoustic Scenes and
Events Workshop (DCASE), Issue 1, 2018
A Simple Fusion of Deep and Shallow Learning for Acoustic Scene Classification
Author(s): Eduardo Fonseca; Rong Gong; Xavier Serra
Published in: Proc. of the Sound and Music Computing Conference, Issue 2,
2018
DOI: 10.5281/zenodo.1422583
8 of 19
Augmented and Participatory Sound and Music Experiences, 2017
The MediaEval 2018 AcousticBrainz Genre Task: Content-based Music Genre Recognition from
Multiple Sources
Author(s): Bogdanov, D., Porter A., Urbano J., Schreiber H.
Published in: MediaEval Workshop, 2018
MediaEval 2018 AcousticBrainz Genre Task: A baseline combining deep feature embeddings across
datasets
Author(s): Oramas, S., Bogdanov D., & Porter A.
Published in: MediaEval Workshop, 2018
Using offline metrics and user behavior analysis to combine multiple systems for music
recommendation
Author(s): Ferraro, A., Bogdanov D., Choi K., Serra X.
Published in: Proc. of the Conference on Recommender Systems (RecSys),
REVEAL Workshop, 2018
Automatic playlist continuation using a hybrid recommender system combining features from text and
audio
Author(s): Ferraro, A., Bogdanov D., Yoon J., Kim K. S., Serra X.
Published in: Proc. of the Conference on Recommender Systems (RecSys),
Workshop on the RecSys Challenge, 2018
9 of 19
Predicting the Perceived Level of Reverberation using Features from Nonlinear Auditory Model
Author(s): Safavi, S., Wang, W., Plumbley, M., Choobbasti, A., and Fazekas, G.
Published in: Proc. of the International Workshop on Semantic Audio and the
Internet of Things (ISAI), in IEEE FRUCT Conference, 2018
Demo of interactions between a performer playing a Smart Mandolin and audience members using
Musical Haptic Wearables
Author(s): Turchet, L., Barthet, M.
Published in: Proc. of the New Interfaces for Musical Expression (NIME), 2018
Live Repurposing of Sounds: MIR Explorations with Personal and Crowdsourced Databases
Author(s): Xambó, A., Roma, G., Lerch, A., Barthet, M., Fazekas, G.
Published in: Proc. of the New Interfaces for Musical Expression (NIME), 2018
Feature Design Using Audio Decomposition for Intelligent Control of the Dynamic Range Compressor
Author(s): Sheng, D., Fazekas, G.
Published in: IEEE International Conference on Acoustics, Speech and Signal
Processing (ICASSP), 2018
Exploring Real-time Visualisations to Support Chord Learning with a Large Music Collection
Author(s): Pauwels, J., Xambó, A., Roma, G., Barthet, M., György Fazekas
Published in: Proc. of the Web Audio Conference (WAC), 2018
Jam with Jamendo: Querying a Large Music Collection by Chords from a Learner’s Perspective
Author(s): Xambó, A., Pauwels, J., Roma, G., Barthet, M., György Fazekas
Published in: Proc. of the 13th International Audio Mostly Conference, 2018
Embodied Interactions with E-Textiles and the Internet of Sounds for Performing Arts
Author(s): Skach, S., Xambó, A., Turchet, L., Stolfi, A., Stewart, R., Barthet, M.
Published in: Proc.of the 12th International Conference on Tangible,
Embedded, and Embodied Interaction, 2018
Skip prediction using boosting trees based on acoustic feature of tracks in sessions
Author(s): Ferraro, A., Bogdanov D., Serra X.
Published in: Proc. of the 12th ACM International Conference on Web Search
and Data Mining, 2019 WSDM Cup Workshop, 2019
Audio Commons: Bringing Creative Commons Audio Content to the Creative Industries
Author(s): Font, Frederic; Brookes, Tim; Fazekas, George; Guerber, Martin; La
Burthe, Amaury; Plans, David; Plumbley, Mark D.; Shaashua, Meir; Wang,
Wenwu; Serra, Xavier
Published in: 61st International Conference: Audio for Games, 2016
A Higher-Dimensional Expansion of Affective Norms for English Terms for Music Tagging
Author(s): Buccoli, M., Zanoni, M., Fazekas, G., Sarti A., Sandler, M.
Published in: Proceedings of the International Society for Music Information
Retrieval Conference (ISMIR), 2016
Exploring Confidence Measures and Their Application in Music Labelling Systems Based on Hidden
Markov Models
Author(s): Pauwels, J., Fazekas, G., Sandler, M.
Published in: Proceedings of the International Society for Music Information
Retrieval Conference (ISMIR), 2017
13 of 19
Author(s): Choi, K., Fazekas, G., Sandler, M., Cho, K.
Published in: Proceedings of the 42nd IEEE International Conference on
Acoustics, Speech, and Signal Processing (ICASSP), 2017
Realising a Layered Digital Library: Exploration and Analysis of the Live Music Archive through Linked
Data
Author(s): Kevin R. Page, Sean Bechhofer, Gyorgy Fazekas, David M. Weigl,
Thomas Wilmering
Published in: 2017 ACM/IEEE Joint Conference on Digital Libraries (JCDL),
2017, Page(s) 1-10
DOI: 10.1109/JCDL.2017.7991563
14 of 19
Author(s): Liang, B., Fazekas, G., Sandler, M.
Published in: Journal of the Audio Engineering Society, Vol. 66, Issue 6, 2018,
ISSN 2471-285X
The Effects of Noisy Labels on Deep Convolutional Neural Networks for Music Tagging
Author(s): Choi, K., Fazekas, G., Sandler, M., Cho, K.
Published in: IEEE Transactions on Emerging Topics in Computational
Intelligence Vol. 2, No. 2, 2018, ISSN 2471-285X
Audio Semantics: Online Chat Communication in Open Band Participatory Music Performances
Author(s): Stolfi, A., Sokolovskis, J. , Gorodscy, F. , Iazzetta, F., Barthet, M.
Published in: Journal of the Audio Engineering Society, 2018, ISSN 2297-198X
15 of 19
Author(s): Pearce, A., Brookes, T., Mason, R.
Published in: Journal of Applied Sciences, 2019, ISSN 2076-3417
From Interactive to Adaptive Mood-Based Music Listening Experiences in Social or Personal Contexts
Author(s): Mathieu Bartet, György Fazekas, Alo Alik, et al
Published in: Journal of the Audio Engineering Society, Issue 64/9, 2016,
Page(s) 673-682, ISSN 1549-4950
DOI: 10.17743/jaes.2016.0042
Other (1)
Datasets
Audio Commons Estimation Results Data for deliverables D4.4, D4.10 and D4.12
16 of 19
Author(s): Font, Frederic
Published in: Zenodo
Audio Commons Ground Truth Data for deliverables D4.4, D4.10 and D4.12
Author(s): Font, Frederic
Published in: Zenodo
17 of 19
Author(s): Pearce, Andy; Martin, Christopher; Brookes, Tim; Mason, Russell
Published in: Zenodo
MTG-Jamendo Dataset
Author(s): Bogdanov, Dmitry; Minz, Won; Tovstogan, Philip; Porter, Alastair
Published in: Zenodo
Software
Permalink: https://cordis.europa.eu/project/id/688382/results
19 of 19