Autor: |
Peter, Silvan David, Cancino-Chacón, Carlos Eduardo, Karystinaios, Emmanouil, Widmer, Gerhard |
Rok vydání: |
2023 |
Předmět: |
|
Zdroj: |
10th International Conference on Digital Libraries for Musicology, November 10, 2023, Milan, Italy |
Druh dokumentu: |
Working Paper |
DOI: |
10.1145/3625135.3625141 |
Popis: |
Generative models of expressive piano performance are usually assessed by comparing their predictions to a reference human performance. A generative algorithm is taken to be better than competing ones if it produces performances that are closer to a human reference performance. However, expert human performers can (and do) interpret music in different ways, making for different possible references, and quantitative closeness is not necessarily aligned with perceptual similarity, raising concerns about the validity of this evaluation approach. In this work, we present a number of experiments that shed light on this problem. Using precisely measured high-quality performances of classical piano music, we carry out a listening test indicating that listeners can sometimes perceive subtle performance difference that go unnoticed under quantitative evaluation. We further present tests that indicate that such evaluation frameworks show a lot of variability in reliability and validity across different reference performances and pieces. We discuss these results and their implications for quantitative evaluation, and hope to foster a critical appreciation of the uncertainties involved in quantitative assessments of such performances within the wider music information retrieval (MIR) community. |
Databáze: |
arXiv |
Externí odkaz: |
|