2
ISMIR Late-Breaking/Demo [Unrefereed] Extended abstract for Late-Breaking/Demo ISMIR 2019 LINKING AND VISUALISING PERFORMANCE DATA AND SEMANTIC MUSIC ENCODINGS IN REAL-TIME David M. Weigl 1 Carlos Cancino-Chacón 2 Martin Bonev 1 Werner Goebl 1 1 University of Music and Performing Arts Vienna, Austria 2 Austrian Research Institute for Artificial Intelligence, Vienna, Austria [email protected], [email protected], [email protected], [email protected] EXTENDED ABSTRACT We present a visualisation interface for real-time performance-to-score alignment based on the Music En- coding and Linked Data (MELD) framework for semantic digital notation [6], employing the Matcher for Alignment of Performance and Score (MAPS), an HMM-based polyphonic score-following system for sym- bolic (MIDI) piano performances. Performance metadata and alignment information is presented to the pianist during the performance in real- time through a web application. This application, implemented using the MELD-clients-core 1 Javascript libraries, presents digital score notation (rendered from MEI encodings using the Verovio 2 engraving tool) augmented with real-time note highlighting corresponding to current performance position, coloured accord- ing to attack velocity (Figure 1). Prior score-following systems have typically associated the performance timeline with pixel positions on score images (e.g., [2]) or semantically sparse MIDI representations (e.g., [3]). More recent approaches have offered support for digital music scores encoded according to the Music Encoding Initiative’s machine- readable MEI XML schema 3 (e.g., ScoreTube [4]). Going beyond these approaches, MAPS offers native MEI support for synchronisation of digital score en- codings with piano performances in real-time, enabling the capture of alignment information at the note level that is musically meaningful to both human performers and software agents. The captured performance MIDI stream—characterising note events according to their pitch, duration, and attack velocity—is aligned with the MEI-encoded score using Linked Data (RDF) structures according to the MELD semantic framework. Thus exposed, the alignment information is made available for immediate reuse, e.g., to review one’s own performance, or to audit and compare different interpreters’ renditions of the same score. By anchoring within the comprehensive musical model of the MEI schema, but lifting the alignment struc- tures themselves into the higher layers of abstraction offered by Linked Data, the score-synchronised per- formances become available for annotation and interlinking within a wider Web of data. This provides a foundation for the enrichment of public-domain music materials available on the Web, both by software pro- cesses, and by musicians, music scholars, and music enthusiasts, as envisioned by the TROMPA project [5] 4 . MAPS is implemented within the context of development of the ACCompanion, an artificial accompaniment system [1] for four-handed piano performance (one human performer with machine accompaniment). Aside from providing feedback to the performer and to interested listeners, the visualisation of performance-to- score alignment is thus also useful as a development tool, enabling alignment issues to be identified and addressed in order to improve the matching performance of the accompaniment system. 1 https://github.com/oerc-music/meld-clients-core 2 https://verovio.org 3 https://music-encoding.org 4 https://trompamusic.eu c David M. Weigl, Carlos Cancino-Chacón, Martin Bonev, Werner Goebl. Licensed under a Creative Commons Attri- bution 4.0 International License (CC BY 4.0). Attribution: David M. Weigl, Carlos Cancino-Chacón, Martin Bonev, Werner Goebl. “Linking and visualising performance data and semantic music encodings in real-time”, Late Breaking/Demo at the 20th International Society for Music Information Retrieval, Delft, The Netherlands, 2019. 1

LINKING AND VISUALISING PERFORMANCE DATA AND SEMANTIC ...€¦ · Sarasúa, and M. van Tilburg. Interweaving and enriching digital mu-sic collections for scholarship, performance,

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: LINKING AND VISUALISING PERFORMANCE DATA AND SEMANTIC ...€¦ · Sarasúa, and M. van Tilburg. Interweaving and enriching digital mu-sic collections for scholarship, performance,

ISM

IRL

ate-

Bre

akin

g/D

emo

[Unr

efer

eed]

Extended abstract for Late-Breaking/Demo ISMIR 2019

LINKING AND VISUALISING PERFORMANCE DATA ANDSEMANTIC MUSIC ENCODINGS IN REAL-TIME

David M. Weigl1 Carlos Cancino-Chacón2 Martin Bonev1 Werner Goebl11 University of Music and Performing Arts Vienna, Austria

2 Austrian Research Institute for Artificial Intelligence, Vienna, [email protected], [email protected], [email protected], [email protected]

EXTENDED ABSTRACT

We present a visualisation interface for real-time performance-to-score alignment based on the Music En-coding and Linked Data (MELD) framework for semantic digital notation [6], employing the Matcher forAlignment of Performance and Score (MAPS), an HMM-based polyphonic score-following system for sym-bolic (MIDI) piano performances.

Performance metadata and alignment information is presented to the pianist during the performance in real-time through a web application. This application, implemented using the MELD-clients-core1 Javascriptlibraries, presents digital score notation (rendered from MEI encodings using the Verovio2 engraving tool)augmented with real-time note highlighting corresponding to current performance position, coloured accord-ing to attack velocity (Figure 1).

Prior score-following systems have typically associated the performance timeline with pixel positions onscore images (e.g., [2]) or semantically sparse MIDI representations (e.g., [3]). More recent approacheshave offered support for digital music scores encoded according to the Music Encoding Initiative’s machine-readable MEI XML schema3 (e.g., ScoreTube [4]).

Going beyond these approaches, MAPS offers native MEI support for synchronisation of digital score en-codings with piano performances in real-time, enabling the capture of alignment information at the note levelthat is musically meaningful to both human performers and software agents. The captured performance MIDIstream—characterising note events according to their pitch, duration, and attack velocity—is aligned withthe MEI-encoded score using Linked Data (RDF) structures according to the MELD semantic framework.Thus exposed, the alignment information is made available for immediate reuse, e.g., to review one’s ownperformance, or to audit and compare different interpreters’ renditions of the same score.

By anchoring within the comprehensive musical model of the MEI schema, but lifting the alignment struc-tures themselves into the higher layers of abstraction offered by Linked Data, the score-synchronised per-formances become available for annotation and interlinking within a wider Web of data. This provides afoundation for the enrichment of public-domain music materials available on the Web, both by software pro-cesses, and by musicians, music scholars, and music enthusiasts, as envisioned by the TROMPA project [5]4.

MAPS is implemented within the context of development of the ACCompanion, an artificial accompanimentsystem [1] for four-handed piano performance (one human performer with machine accompaniment). Asidefrom providing feedback to the performer and to interested listeners, the visualisation of performance-to-score alignment is thus also useful as a development tool, enabling alignment issues to be identified andaddressed in order to improve the matching performance of the accompaniment system.

1https://github.com/oerc-music/meld-clients-core2https://verovio.org3https://music-encoding.org4https://trompamusic.eu

c© David M. Weigl, Carlos Cancino-Chacón, Martin Bonev, Werner Goebl. Licensed under a Creative Commons Attri-bution 4.0 International License (CC BY 4.0). Attribution: David M. Weigl, Carlos Cancino-Chacón, Martin Bonev, Werner Goebl.“Linking and visualising performance data and semantic music encodings in real-time”, Late Breaking/Demo at the 20th InternationalSociety for Music Information Retrieval, Delft, The Netherlands, 2019.

1

Page 2: LINKING AND VISUALISING PERFORMANCE DATA AND SEMANTIC ...€¦ · Sarasúa, and M. van Tilburg. Interweaving and enriching digital mu-sic collections for scholarship, performance,

ISM

IRL

ate-

Bre

akin

g/D

emo

[Unr

efer

eed]

Extended abstract for Late-Breaking/Demo ISMIR 2019

Figure 1. During the pianist’s performance, the score-following system highlights individual notes in real-time, coloured according to attack velocity from pianissimo (light yellow) to fortissimo (dark red). Alignmentinformation is published as Linked Data, available for immediate reuse by musicians, music scholars, andmusic enthusiasts, as well as machine agents (e.g., for review, analysis, annotation, and comparison).

ACKNOWLEDGMENTS

This project has received funding from the Austrian Science Fund (FWF P 29427 CoCreate) and fromthe European Union’s Horizon 2020 research and innovation programme (under grant agreements H2020-EU.3.6.3.1 No. 770376 TROMPA and No. 670035 Con Espressione by the European Research Council).

REFERENCES

[1] Carlos Cancino-Chacón, Martin Bonev, Amaury Durand, Maarten Grachten, Andreas Arzt, LauraBishop, Werner Goebl, and Gerhard Widmer. The ACCompanion v0.1: An Expressive AccompanimentSystem. In Proceedings of the Late Breaking/ Demo Session, ISMIR, 2017.

[2] Matthias Dorfer, Andreas Arzt, and Gerhard Widmer. Learning audio–sheet music correspondences forscore identification and offline alignment. In Proc. ISMIR, pages 115–122, 2017.

[3] Eita Nakamura, Kazuyoshi Yoshii, and Haruhiro Katayose. Performance error detection and post-processing for fast and accurate symbolic music alignment. In Proc. ISMIR, pages 347–353, 2017.

[4] Simon Waloschek, Axel Berndt, Aristotelis Hadjakos, and Alexander Pacha. A showcase of new MEItools: Half-day workshop at the Music Encoding Conference, 2019. https://music-encoding.org/downloads/MEC2019_WS_Showcase_MEI_Tools.pdf.

[5] D. M. Weigl, W. Goebl, T. Crawford, A. Gkiokas, N. F. Gutierrez, A. Porter, P. Santos, C. Karreman,I. Vroomen, C. C. S. Liem, Á. Sarasúa, and M. van Tilburg. Interweaving and enriching digital mu-sic collections for scholarship, performance, and enjoyment. In Proc. 6th International Conference onDigital Libraries for Musicology. (Accepted). ACM, 2019.

[6] D. M. Weigl and K. R. Page. A framework for distributed semantic annotation of musical score: “Take itto the bridge!”. In Proc. ISMIR, pages 221–228, 2017.

2