SunoCaps: A novel dataset of text-prompt based AI-generated music with emotion annotations.
Autor: | Civit M; Department of Communication and Education, Universidad Loyola Andalucía. Av. de las Universidades s/n. 41704 Sevilla, Spain.; LEAD - CNRS UMR5022 Université Bourgogne Institut Marey - I3M, 64 rue de Sully, Dijon 21000, France.; Universidad de Sevilla, ETS Ingeniería Informática, Avda. Reina Mercedes s/n, Seville 41012, Spain., Drai-Zerbib V; LEAD - CNRS UMR5022 Université Bourgogne Institut Marey - I3M, 64 rue de Sully, Dijon 21000, France., Lizcano D; Universidad a Distancia de Madrid, Carretera de La Coruña, KM.38,500 Vía de Servicio, n° 15, Collado Villalba, Madrid 28400, Spain., Escalona MJ; Universidad de Sevilla, ETS Ingeniería Informática, Avda. Reina Mercedes s/n, Seville 41012, Spain. |
---|---|
Jazyk: | angličtina |
Zdroj: | Data in brief [Data Brief] 2024 Jul 18; Vol. 55, pp. 110743. Date of Electronic Publication: 2024 Jul 18 (Print Publication: 2024). |
DOI: | 10.1016/j.dib.2024.110743 |
Abstrakt: | The SunoCaps dataset aims to provide an innovative contribution to music data. Expert description of human-made musical pieces, from the widely used MusicCaps dataset, are used as prompts for generating complete songs for this dataset. This Automatic Music Generation is done with the state-of-the-art Suno generator of audio-based music. A subset of 64 pieces from MusicCaps is currently included, with a total of 256 generated entries. This total stems from generating four different variations for each human piece; two versions based on the original caption and two versions based on the original aspect description. As an AI-generated music dataset, SunoCaps also includes expert-based information on prompt alignment, with the main differences between prompt and final generation annotated. Furthermore, annotations describing the main discrete emotions induced by the piece. This dataset can have an array of implementations, such as creating and improving music generation validation tools, training systems for multi-layered architectures and the optimization of music emotion estimation systems. Competing Interests: The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. (© 2024 The Authors.) |
Databáze: | MEDLINE |
Externí odkaz: |